var/home/core/zuul-output/0000755000175000017500000000000015154577340014540 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015154604656015505 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000267322315154604571020275 0ustar corecorey ikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ ?KEڤ펯_ˎ6_o#oVݏKf핷ox[o8W5^|1Fr_?c^*߶E٬:rv筼ح_y~̎+\/_p/Bj^ֻ]Eo^O/(_/V?,<']_kmN:`Si{ C2i1Gdē _%Kٻւ(Ĩ$#TLX h~lys%v6:SFA֗f΀QU%Ty2Kݙ$ӎ;IXN :7sL0x.`6)ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&c(޸0"$5ڪҾη*t:%?vEmO5tqÜ3Cyu '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=dC' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGtGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YN%:ò6PT:”QVay 77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&?lm$K/$s_. WM]̍"W%`lO2-"ew@E=Y]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"EkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nWppH? 8>X+m7_Z`V j[ s3nϏT=1:T <= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@n|Vo|6 8~J[,o%l%!%tyNO}}=ʬ-'vlQ]m"ifӠ1˟ud9)˔~BѤ]һS8]uBi( Ql{]UcLxٻa,2r(#'CDd2݄kTxn@v7^58þ Ţ&VY+yn~F8I !6WB3C%X)ybLFB%X2U6vw8uUF+X|YukXxVO(+gIQp؎Z{TcR@MSRδ~+1æ|mq՗5$B᲋eY(|*磎\Dži`dZe j'V!Mu@ KV{XץF .Jg< ƜINs:b zĄu3=Az4 u5'og^s7`Rzu-anOIq;6z( rx߅ eu;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKdٻҦ62L0ډ"ܺ_z9JNȯ=@oUhN/ٟPE[/Y5d{zrBܖ6Hlc "mKv~[uLU4lZ;xEN'oI㤛rP*jC# 6@dmHg1$ʇȠh#CBΤ{sTQ{%w)7@y1K^ ].Y$46[B-3%OONw8d`Q4d$x0t8@t]y1T\YAidtxBG:pɨU<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;*3b\N7dYܞLcn3rnNd8"is"1- ޑܧdd?'xc 73\+s$N9tf{w5n%՘#TI{h5pfIWGev.F_{!7*qW}ӼتY&7!5Y'3KtpwU`<d5J{@ݜ P?&pWt|CX6|,9:N\8|ҳ uE UIL`F &ni?4~j3?m-Kھnۅm''J[/m$g4rfrvRޟ/O_·ogg]fyrd-i-Iv#GL`,Oȃ1F\$' )䉳yg=#6c+#  =J`xV,)ޖ,3~JPͪm|$oV1yU<̐t6 T m^ [IgINJ\Оf*Z"I)+>n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}q3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwGԥ1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[t4*6E7qP'cts3tZ3lPuV-]fz:햘s >˧Byd}gs9QN. /ӦxbHHAni5(~p>/O0vEWZ nY3 cU $O,iLacoW1/W=-kqb>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?hV#ہll/ؽnA(ȱbAj>C9O n6HNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2w⠪R/r| w,?VMqܙ7;qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z]gQ)Bی:D`W&jDk\7XD&?Y\9ȢG:${1`+i n8=%Ml%İȖb7AޗuV3A7ำqE*\qb'YpuHƩҬV nm=Ɂ-2=|5ʹ zi ' ׹U>8bK0%V\ t!Lku`+]c0h&)IVC)p| QUA:]XL/2La[Xѓ F;/-rtx-rei0hE˝ݸDt#{I} `v;jUvK S x1Q2XU&6k&lE"} Q\E)+u>.,SzbQ!g:l0r5aI`"Ǒm O\B!,ZDbjKM%q%Em(>Hm 2z=Eh^&hBk X%t>g:Y #)#vǷOV't d1 =_SEp+%L1OUaY쎹aZNnDZ6fV{r&ȑ|X!|i*FJT+gj׾,$'qg%HWc\4@'@—>9V*E :lw)e6;KK{s`>3X: P/%d1ؑHͦ4;W\hx锎vgqcU!}xF^jc5?7Ua,X nʬ^Cv'A$ƝKA`d;_/EZ~'*"ȜH*Duƽ˳bKg^raͭ̍*tPu*9bJ_ ;3It+v;3O'CX}k:U{⧘pvzz0V Y3'Dco\:^dnJF7a)AH v_§gbȩ<+S%EasUNfB7™:%GY \LXg3۾4\.?}f kj· dM[CaVۿ$XD'QǛU>UݸoRR?x^TE.1߬VwխmLaF݄",Uy%ífz,/o/Z^]ݖF\\UR7򱺹...m/~q[ /7n!7xB[)9nI [GۿsH\ow!>66}եl?|i [%۾s& Z&el-ɬeb.E)բA l1O,dE>-KjLOgeΏe|Bf".ax)֒t0E)J\8ʁ,Gulʂ+lh)6tqd!eó5d ¢ku|M"kP-&ђ5h ^pN0[|B>+q"/[ڲ&6!%<@fpѻKQ31pxFP>TU?!$VQ`Rc1wM "U8V15> =҆#xɮ}U`۸ہt=|X!~Pu(UeS@%Nb:.SZ1d!~\<}LY aBRJ@ѥuȑz.# 3tl7 ]وb Xnݔ[TN1|ttc‡-5=VrPhE0Ǐ}Wd|\aD;(;Ha.]1-{s1`HbKV$n}Z+sz'ʀ*E%N3o2c06JZW?V g>ed\)g.C]pj|4逜*@ nBID f"!!*7kS4޷V+8弔*A19`RI/Hй qPq3TY'퀜+/Ĥ'cp2\1: 0mtH,.7>\hSؗ΀ѩ آSNEYdEcaLF&"FhQ|![gIK v~,Jc%+8[dI368fp*CDrc3k.2WM:UbX[cO;R`RA]d+w!e rr솜[/V`+@;Τ`5d0ϕ_Lع`C"cK>JG.}Ε00e>& 2䯫vNj31c$ i '2Sn-51Y}rE~b>|Ď6Oj~ebIapul9| 3QtUqSCxTD7U9/nq.JYCtuc nrCtVDƖϧ;INOKx%'t+sFUJq:ǫf!NRT1D(3.8Q;І?O+JL0SU%jfˬ1lމZ|VA/.ȍȱh M-r ~[0AG꠭y*8D*-Rz_z{/S[*"꫒?`a;N6uilLn<Yllmb rY״͆jqTI!j.Pٱh s!:W_´KxA|Hk1nE6=W|$O -{]1Ak$ ѫQ6Plp;3F$RveL l5`:~@c>q,7}VE-Q8W70up˳ A¦g/OEU:غA>?=CۣPqȅlW11/$f*0@б 2Dݘrt +qrx!8 J&[V =͋A,z`S,J|L/vrʑ=}IhM4fG(Ȋ1{TT%41Oa'$ ]o۸W|_*[OJ*iw78M>bP"eFU=d;dNsZ?p49xjDHU'[JP&o$8Tš*\d|AA-Sa#8T(9(E7Orl!CU \.ƋTxB 5-fc*aVܞ%~C9Tp!_2j 2'R<*CUE䱚v:XxZ7qy[A?TUmf4QNz2(jrir"qF g% 8 ajvY}B`.ro qsKT^ @tJʰ񔔭))[arCe[V ) l+&/;_W'-7~csuFD eqxwzY\fb[agm3X#jʈC Wfܳ }m1ke= L&9~*m{y!\0d9Ϋ|ӣ ZĈI&Xny:x''J#ֿ{6S ["kW#OM=߳ T:(Gzd1[^RHKBoL/aJ]_K,Ď'볉J we,bkQՊL*$DJW2xfpOa)1(kx=^.QkJ\'4 kŖqp暶aka`ysY$$9h'TF4d3-Qv]*qHQ\.Z~sD*(Te*cN" 꺤E#)eibKa`=r&엔Fb(AyE B P9(<1^S?sQ Z Us٤ >>R2]:|'FU:x4|4z=LդU5iW RoIU\.қWT'D b$rjN-gKGʚlMJu)8$ZWoȊ:m7xxsx׵gkdӅu\VvwɟwLߖ4S+1`g@S1SA`˥N>HDIo(Y:83 D%S~:])RԷ4-13dz8 !z.&I4=>Á Ŵ%>@).qw"0~l'~Se-i箁WyQJrwF7nÆ v'`gI@) %Ĩ{S#~K9wa'' FG',Jwɷ s$E }5?nN D i]re9`(0Vaoۅ!"\qI3pQ`ml,(DL[QMQ  bo{wv,?Gww1{*s) G7~v@j CJ.(!"u~;Ã6(]&4s_ Kebww]5ch{9U־f .ohCDM W*);kyQf,D7_{go[%BqNrn.`tɾJxpXۍrko|!b9aL헦R8Q*]/YGh֝^+D;,3[(0sκt: {CS\ܴӧ-PP[ջ)Ǒ)Q=.)[(&ϛj 7( 5\ $ChQ̬pHPDmd㣓޶)kqzM#ъCػix,_#C,Ѥ(e<0RIIs&֥`8(Rk 3!gH4m2Q/n[] Pc9♪*u/&ߴVG!U[ϲ`;pCILi:ɆK"6>봤 74$&8,kNOXxsOvbZ}Ww_vXtLBTp~re}Nϥ[sܭ9#rI\+m l ꍵx1r1Un(3+" {& ƈ*^h?ڈdmL bnA8um.o 8ıK,T Ϡ^ )nG\ՄZ!0W jQ3e̩\V" ~zHydCI`"˘9mG djلsԈ+ \GMY X!ЬPT)UPG'Ic} C{5,K`X.ْrMY {Z"Ym0^TdXɶt nJn%+ѽL:gogóA2(:|#jZsSK['O׭!f,J*3=sDUzG(ZBE :kZJ8p5Nn_L2T1 a9P@ssSHs]/0c._q^(N(Z t9% eUbJf%1H쮆g~[ų툫 ߲G0L!Pnm*7Ԅ,XFl6%S_*zOӭb6%N6F:+E ;$[eGs1۹LpƐX[N]S6tR8:ؕg bzus+h!H=ܖ`Omk dZ,PZFMåK{p ,{MpgI,M\UW)L{ZWD }@³trFs62EZ1KR"XgͣҐYfK,O  ,=m6JƁa! ;nm^p(lMJbY~Q3{{)A<AmRu_[JT[=H0l2D^:/J]5(TW}-Xx5}F}"Ky,2 -ة:hW365_w--23XU4(?0Ta]U+s.}{޶T4Ds1qa,ܱiPu/U)<s%ܶi$wnnfh[Q$Pb9q8S$ q.zvW49 AZop_nVC ul=wt'Un2 6ҹrYVCeP-C۵UD%=tu$)y[B(lI5iw|Oޖڸ~SAN;Ht^ :p@Wd=oIKZbi`3y]tuZu,h%&T/jDnzIx~"C]7@"UGԵ,d*g(|ݭGHs ȇ$_R9Z%quQL]k-םV[dR%kja57(l0lvwCym=2`f2U'L7=i O$"6:!&hLpf+4+ d 8}v'{ ߕ}1i'ܿBg)"& k4!V3b!&,;cئwƈf3HAla%{7'؞M=&Ilzo岪۹XwjCs/zi#Ns_a ;?8bq=L6N~ALSP71i~SNMwb 5qD` ) Ch0E/ED8!.3[qǙwK\V5=o1;| `4`6qk $ɤ#!ÅaFųwwNgAA,LÚ2Fߟ*~vD2Y>bj#CAV?{W6B{W;$E ٴ v%"%9ֶ|4W~3$S:Ҩbc[ f3#KlvS9>bn{G=i!wJOZky[٠ח&z{pCVoZAd;^Fk+k3T3vo}>a/sg ˈ|ŋ?*Wn ;Vee}"J^a,=O量U\+6t&'&ļJ)۟<>ӷg׻ӃW? ow/fP<+ gױ[ΞY~cH)^8{|;{!4'z!"m#_WJ o'I |Gs{IѾ',\*taޏP4Y1XBa?Rn=ǥMFG^Ы8^Oݞ էlဇJ"Nwa#SoJyo=$~|K>];~R !}nvEoοV mKCC9NHkLJ_T{"% n#.3}@WXkoG8܃뼟_?0^E^8$+}gK2#Iv5DӐ޻={U>V(>/&0#) TFG4 y!" 2 v0 ea)J iXeB / J- 0`]&o҉|35Qﹲ*2p1DvݐE<)ȸgĹ5O@WE6s+oD[?0{ 5 zJ8Vg&fW0NrŊAh5G*$s\OS o XF~h0;ysh/15 9fGqIL59< ?p9!2XelߩG^<BQ-51bi12caș%|sРy <} , HROZV:Dq$%X|+ꪚAKaE\.2`X?eptx~fW{Wi,*0>`@R],(A(_$ͨ#6&$,-tD2YShEW.Z䧂S*5Yeq"3/"Jf9|r*a9[K2E9D/Ю e.yU[ד}^Ոv5 ܪEԵ|FWyI(KE8UF^vUpaWA-Osg{w-t!c Vtw0e Y u=/b!ЗAݐÓe4/^F+X?`H_7!A.'7%^:S*pXV{CPq6ˋJ@\&T,BuQFG8{h28X,^:,7_]^ ";A;`T](R.@ [ EeOD1F~%L-FX4ϲ)sYziH`8`P$ |؈h%~ 5/3aW#U? O=av*V_=mÒ㲺J>n^Zmrjm TpΆ+ݔj[4<y{ws}Oki/Xрylĵ8߭Ǿ4#x>.7U[c!nڰ @򐇂rw"Vk8^#VZ&&ucȲ/2KWD|^@ $&Uݨdq)ɣq ]P8I$|~qZ"&).h0M028|~Vݥ,j˃.<d) yY %P&@(۝P8B eѥl>eT^eVSnvN=}l>sw]ަ@,ߝV8R); uw ݝPq$݁P&T@؝P8B ;*ۄ*G| rB6z=Pz; w ߝPq$߁РMh#4x mB w'4| K)KBڳj4ϯDwAE j6߲dϔMIx:8r=Y^%I.r<>܃{Rj7!J%3+JCeh]UL+&T MrĐ0b9=96do˺)tBŤȯReޞa(@rF9%5Wc[N3+6r/j93h!-[yJRZPR8 }yRLrt8 =n%u0+ϫ>_6!HSO N3uܙVִh>u^q=Ϋ8/  9,Vtelt^C0*+͔! `WПY^Sl/Ba԰Ye:xW+EW'۰ %@< .Nr p',Ɠq4Z#LuY#}Lx4o͆/ƈ`"UvT4G@wުP'2,"X/;N/ݫhL@DEL Ho4; 4WQLl+,M]n&pL﹅z{_z=$ Y}.ڽlY5n81707H9H4"c #x +6X"S$;9j?>8M2 iUaCN?ܗsESgՊ>OQy Cl)BC'gg[˯[d2I{wnV{~m@xn[ȈV]÷L\9M[d-9kq2;I1`Ɇ.ʁq^! e~J:YCmr @N "(_bw|T=?aLYA֯J2lVj1j]S}Q<ݖt<ҕi˭E݀>U/jAX;w;tPY ǥ 3Jz+I3`FOR|4l2fgG4{n6=PUfHqϱܰPESĤ:C~GZPL}00ϖ5VgɄS m>Qp^j)RKuE"x}iRMt(]FC^!:?>r=6LG2U)U @A" PVUf$YjvOճ U|+exnF6wϰ$`s}(Ai$.VS<ʱse܆%4@6TYj_ =o du?g50e'3P77tb].Di0l|n57Q=UGY8|Pj|_EYdu_nS\%җtjȣ|[qfVDsPkImE>dv=nU7F465;h.Y@[85mB񈬔<* ZJ"R15!=C0kͮ _3*%ɘuN lIׄ 6F7i|Š1_i6ZghA|vM|VG\&48a>XF/ {=qUbL,-KԞfg= Ԇ`]Yo[r+yȓ*}1rdW%S}HJgĉ]{phx>LvaO춤/AIf,)" c֫~̦jχA$^c ~OcRWw}^j}vƂg}W?ھV[}x+io5?BzR')'PۇviU[2=1-1-%d~30!m'B_%M]}cW2ICITZe뽀gn=2gXE |E}KCmܴQ[[j!IweއYx~Ɠ2->g|I[ 9FJ(1k>:bp9 !jECöL-6ӜiC4zxTPRf)0I'qc{hX]y *{TDY"eXl[*)(UjPx\1ο SE5 (ƪk+׵UE]9_" E JN0M|`Iؖz`Au{M2Mf:*pF% PCH7-p G. wE& /jDTOj)LDBtr/>}v}pއϻݕ]Ex8$ܯ~c z=hjIˤPC=ec bԟ-KJ%|yV Qj̴LSXv25nA3͍&pihQ$l RpJLD Nct[\4KLtbtQP\*FyssNR 1(A*aѢG5Pʢ1+ )c!0IS(Tb2ωFtz[8,Zs05XxޱʅBXH:1]?1$]ă4A'U1)9v0[>*p=(ccÅ^GDIa&LS{88O$JW|a,љA4)xbL4}Q7Dbۤ{cX1B ߖ+A/k֪Lt:1MGRj#l r,*@N).9LR汃-dg-H76P1*&Hk52p*W8Κ]?$Hw+i]<7n>ŖZ9-[nW|YNˌ@[NǔM M璁_q$X?m?vj+8tp~^1?x6WF)9ZR߆n,+^JwR$ vl`ǹ@v5؝9כ!Sߧk܅ E;ַ׻[ȡ\b4qߪx]23z09xVc)\6Ft[2 0^S]~h7fy(dŘ"#*^@41I'8nE~ZwqsnH$^{'׃֡W,(-D"^B4 q<A҉J><>Xpc,Ĺ,W3C"^{y=J,$fDXn]XUIE۔wЀҧbݲvnaRĻCZa箾(/KtwIH?7،\56 %$V(NQVծӐ&J.zwe_S֢?MPvst,zQљC61މi^?W]md{2=&(6v^eҋ<bH.<`a"eC|]Ѥ,Y. J<!6v7*TJX(Np|h肣]VūMjaq])T.@3p.f$qnPEQl^5[-}n5~NO"Tބnw]p̂e:Vߦo9pc t>jT"ImIr`e 96*I57. );eeQp M\ nQ53JAu7Պ3VέMWjS>8ب;Ng' ? SҗGl-Vbu*_X/yKZŘ֖k\C=q%}#UqF}jMiJ,ĞÀFKYR,+>mբ('Zoਖ/:#nR'kq^r1S#owk@E-|1Yj ]墲]/qUrtǎgc]$hI۴Y=`;a`u]dlHXε7ѣ?C)<ڠ֐)V"sˉ=$a9L[Ad%onjΈiwe@HUcDޗP^%^>+ p׬Ѽh3kuZJm$VI¼3s;4  Q`hl=DzʞD<Icvp'Z6еdhiK"{=ssꠒ7 ܔbh7C*~8E Q$ KϵKIײ'qwЇC/(+)eS`އg"r N/}Z7]p4lQx!X8]?>>tъEqg< ]} ;qɑ`JZG=;~@Q齝/zKp쟺ýQ@Us)\fE5?ֱ8 JXW*WfQql+%>YّSnQg҂ݗ)E<|}'+ӧ.δ\j:7𪄤aצXl!@`VnJjo]MkM/б-][\ڀ*5(ۂ g5 6{GG\vB6GJ ɜ | Jn?zL/Gzc7/lle?ߚKd w9%c>׫ԧٍyvn| Qo]qlAl\sܚKN' nn,nEkF#A4cFg(N1tmF%}Beǂ"'[#l}ss*%gs\κTQqoҪQNJ9"Hj{s4ߵ'̉9saIڇuʨ+Tt07R$̚ngkB5X MI'৭=N2iK1IrsD~΍nQ#D5s6"2dMӢ siB_P %peq7 rGߐR%zR٥a3r) eJ82 y">A`e5x>y-9EW=CC6I 4GwͯreN#Q;qUZc'yLW [h7ca{SLwZgAή-9uegDu<&[9r_Vf1ux>H= Y뗇KQw݄7OC??c{ :60?'IWӸN1^h9K7A.=ݴj"UY8jȸ35TWy@L\DhO0boP*~^VE=\~ǘJNop_`* |joX߫w\|]"$i<&"\:@IKswВGWuК鎫0ȓhXʶ腮NR(7?`53XN!WKXDh` Ͻvhcy F=k_]ݱ:)^Y_-D-D%\,М`Yҩ3>y8n3ϓG [JsNAsCer 89^Ao,Z3שOyz.>: c0h+l\>oF٭ 3{݇oJ~Uq?<naz$nqu|?>VZR;aC; 7`Cz AjgI6rSȩ'+߅/|%KLG/s"k֤ju-g )}%அ]-7h.fph=]hA(u2{cr<+tJL6֫>DUU>̿gVjy;˫`hF'ӕ^/Gh)L\&`]^02K*4dLJ[Y(2=33Y~Z/ x㳔+(Gd2N0uمuƕ|i [W%rhe VI t?uڝr*vrg $MM¾MG\9_J{L l6c0yS[_.p'FVq (ys]agyy^ଝ۳m ["NUWx-|K-׋Wۃ&i߀ۉM>ӕ2TUoNƊcQ ŗC cc&:8QO6$ծK،YpxyJ'epxVqt29L6xg_qb+u7n6Q7=+FܣӴY9wpgZRj??==Lu^ct31Mʜ%nL ȥ5%sJ숅VUliJH"Ȇn)sL7&=އRyS8)(gJj}V ,-A D"-Ai ՘e7ZbcϦPTrxQd\_UWTbWTq,iPI%UA%C7@1T)ۍᘏjЊlXFFh#5H(-% 5JhUPaPW3藺 0K," \&-iA $ QiEo8MTx:fF C%Q AuҰP'!( ʧ`| Zhv0Ð?h}9oAH,eu&W| +]H Rqpݞ?Ł!Zƣ )b;Ӿh1:5^GR-[Utz|bQ@-7NEA<0B?YU>`CJNm[ٷvE_pg~mo8:IIgQ`f5"Ey!.' NfM~WᲩ/΢wU k91'g FJO^S^aɬ9,pfUK9GܙI;/9żEN8:c}V9: GǿMJxjӁ`*Dc&1@-[c}:fO:zGpN2,c|Tjyh~.wk<Ju)>#[#_nY-dⅠi ~zSM>z'P[k(hmu2=}*lY˜/ۯ㕓,Á8 kD(;[IY,x.oն} rpeQq3^Ĭ;e[^gaV՛s^ .+70kL|"'j=:x%DCS9s㴹Gq, zٙ ~gc::xrl96~8GYa1 akvß'v<'PFJ&/M iI@\qp)Øc`[t"*#gf^;Fg~(syyn?bnGgXn OzQgyC}Ο.  Ru4R*J12b5瘇 lHSpz=N2|A2RORc| VC v`XM_DeCUqWu13j/[=H"C "Ḳq\QQcnGjח Ai~ۯeۯn08•>VLVǙG B8j _0o XȂd`"3,0VxYݎj?٦!49Tjg2!?s#):tjg3r =ljg3jajg};hwjg};C4&ZKtĔvY:~ }a67]t}2k!4 Y1n.8BLC~ˀIc*r1W݂) DZ6ufKj?zR݀ks1é" ~۵c"˸7m3n$iVN8:f!J.அ]`^=0@M^>QUCku>߆.5PgofUuaڊ0fR<_iMn}o~~w7i5S1q|eDV 7ӀCX\ImL)$\o 3W󀮊Kn$W P!I5j*su૛plyV UdQg\'wŸIT^x/J*bLBdxwy:-5{9=N$H G9V *` k4 1\xx&,[)nIj |uRyy5R?`W;1XD0Íq1"G軕y}5ϥ.XoȜ0I2 /%\; Mv8lI 8;n؟pҴM@n, RR ,G4($,^G[MiH`+yAс+xJb0ïg7kP<`Q%M᝷ٵAa1BJ1Ęs$ R@nsQ9[Y64HuLv̊(%I9g"A!a!p"DD˜kx|DRG%|T@|3RQG; RflS8BTND;Gy-- " ("譐2k̭\TQ(Ӈrй:O(\FkO)3n$cIUNX>REL<ƷG)I(Q~qx@4jJ+(ow̻0zQ_1٤G1k隵38}]dޞYRx]鷧JXEZ,3 > \T+ç_:f/NC` р/6Cp)#Éep*!T F zn OJ c (ě* DkMFI,(cRR5SQY$As19έ"Ċٴ@$wYMtWjn$&u+|5 aT{ikCh3)iH[b"A& 9 f:č>, \>ڌ#0l.߈`Ie3r] zRppo% :C1uʩԂ+0 z~$s @sGBdp3ga&PRi PΤZ_ Vr\:i vsS}F "5fiŨɅ3S DH,8` 9h%w-mI`c%M9ZlI6U` gz-)95߯zHICQPLï]y͈j Grd3C #U(29Z`miu2>A̦ʷlzUkjU[V! $3-Zءi͢ poϢΘ-R\_֑'y^ +M`S'z$U_B; <\@']?KD!综Ɣil.#s* sd9H̅QH SCiC *| :y9>vtWd0Wg;дg44kj}X mA ?~{P֔ԥQb 1 vieJXv *]9eZҕE*]9gڣCTVF[447ɓUxp-h!&ήӑhu_k!4Li}i2EVd"/H Dϊ.l!Jȇ2R ?DaES,MNPPnJa^SvUZ#guQ5R9SU{JvMAp~p3ls`3k h}744~dSeZ4U9K'\ECcAKʼ ָkTNPkqͩB? 禪 ?_ބ07ս;+ąk!1<\D5kώڂ1π-q%eG7:GtY`!+2zz6^osJ  9SeDN-67~5Q'VCao:ߑoRDS&j$<ʵl~uLO,6u0#DM"P804 OXE\-(.OU5g]L?7M7:=Ɵ8ϿZx0( ( 1\wGqsYOoLX#*UhLB$?di ԉId:M!kDJ& T8oC`HÄ 5PxG 'ФP[mp'Dk6au[6jCmxi+04N0$:kd>=Y+ 0ᚴ:lTEZxrOi-oNYS¤1}WO:+/{q+X/2|p|VFU;mA+eqcUE9r|7%yɭt@XZDTS3=IPrWDWzƯ ji53 C au9:t,Fmٻߧ.h'藽}Nؑv?e}zq0 e- ^yV\Ɵ>}vF8˖Rs*sYZ5^glM-x^.|>Yե콩qq^ɟ4ކ]`g3&廳]pw*ox={y}-/{/.IT#@ރzooo#ܓJu+ָ(;-z=Z.Ha{T5sjKS94'Degh|~ڻp.N3L^9542ٓ=d^zZ?<}U˓v4:AZل|1}L+Svr\Л`{A ZnFn[CVh% ubt-KŸ~ i|Siy}urR4/Wlp(G5Y3_' /af 纤p60Ѣ^~TzE׷zsv|O6?,_;E淤1嗿vA〰LE v__Wm6OS慛oMA5)ӟ:?8?f+ W2K{z1ĵ}AEfmN$oָc:"BEiv[#|3} /*Z qQgޯ3Avv~4!T(VOvY*~.׎O&Sa,;rʯ7 }MC4aa"WVʣDY(P͘&%uW3M4)ijN߼i(׺zcY@Lzo/8қ? R:Z 661v!W5.xyS^d, S'!e)"]Z ^rd{ɞ/e= {&CH<\@]gܻw(֏h #Ʊ;Ipmџ*kwElR/=DKR&%b%u䤜T4SWl%vjB}>Xq;1BMwŵ< XxŀZ`#\k|[@$#ԦYu1^sc<\XurH;{gϐgd0ew sO%u$Am|IE.SkH L &`*4IϤYhC4h; 0џjn'{ix#|r0N; *(/J#mHBzG#Z U6=\E;_7opp:!"ѲXKtpZt7+PoU! `8>&M""!jR,pv,hmyyb3Epb?l F؋#V2cp`}Ol H[k"Vx޾{Q6{,lQ6ol^3EcPNh!w"o oIg72 fOa6d)iQ&4BRKArX]]c~[?90Tr h&`+<Nl<9 <`=̂,Ҏ4"W<F5Ȝ)H/[T5͌(QBeQY 9MzP[Um҃X=x8֓X%c xJF4Pd|]MSCuj2XINj,I-FY@]^̋"*S}G ]Цe܎bǑKx y^kAEZЄ 5->)ZZx%kZ:97/|FKeNtEʂ0)|ڋ'MbC[x=cPZWPt qJ49ĝI!.aӲǻyyXO]t $ 9Tq7ܑ3WP1n^j vk."֕JQ=s(=ImL,ԣJ5/'a+r;nȊorw< #lT݌;ʃWyh҂X[1);@F$>\~q_nkY0ErnI&LL:]j1ϝîqXl.ѭo*\G?鵯19QZc{Ov>vlj3 yhb&~#gm7 =x8:8k.q U =Ք[M!Б 6zI~rn\ O9RlulG ]y[1nA&.hyfydl[Ѐէ(&.jl>uէnQ} *F3;Y~է.Sq9fN^X^u_7/{W|/TVgŧq7 /Z՗k|{N5^B,SXΨm=r3X'X0a\i)C'S5#ydkNbxv2婦n.H_'\nyϔU𾉔Ӌ~kmw_Zf{RGۑOj/>o\~(>]C* <1ȯ5=@PS;sM׍!?o D1dc,p`xTsaQMVg<^k}Ue7te<FH;<:Tݩ'ޖU]9i*hk<S]BC^Tݵ Pi + [mS?vӇžܮWiܕ6Kw}E_C~UtCk.c'֗F9b[cy?/G4 gPXrއu>Pm^ێg`rc]VbMڲnKbWi3°` Â]_[~! 7K~-XZџ[n!rw˱z&zثN1{IeR[&5Fe oLeՊ\BeE`X~YUNn Iaka[M,]$Fd=Im$M-U[WM"Gu)IIr429:K[#Cy!5*zM$Fk ֧#^R8`d>b ~mݛ#{UU|q~zuB'k23כ#zWH&nov{'s{}GΩPkj^іt{=L%SfsԻn[}h/̞cW W&uU&ڳS7P>Um $ k*&֮-ױuS_~'j=sCJ,(55vXjwKDPX/,u`1G~ԑX;֡|6n P/F4s0&@`u6Tw]3y0PUi vԽ ۰5VZeVG9#K9k+\a#AX ߽˲9ݽF׎^>(7a'g, J8E(YX;ȾΔ0BI8bѲ7AՈ1JG̅#`l~9؝ñ8zq6Σ$P (#3ݏt  3;;+&!+81,@I62BCCWqL ; I%. ]VaƑ!#[~E*T"B@Ӡ QF#d]u!."BMƶZ֗o>IF#}9 b\ T9xY}[ki.uI5A_R:Ml}A$:Ƞ$9Iqtҗ]w! uk F#} 9IdB}"')Z% uq&8e2lXz;>d%M8 ωY2%,鰽b ni ),628;RX**̒RUoz[ϚFцfSgwy;r9w+qE՘BﳬۢMiKA1#S6YOEkE)~+0(mhBa#-3垢\R r{ܳW=_lV8ky< `bc6tq3]44/<0}r8PoB ^̇/(K0P3_v×1ҒEPYP@2h7hd>3ek`PgvϏzN`-mT>4mŏ7̛yc2o͛Gq )1)D5tI ۘms^DSaKSryF44td:'s0Pdd60"8-.z.DRaA y-{bQTSV5]EQ/q6_E\tU7O5hd6 }ޭڦv`ޭv0Go=l!'NވloIF(( ,R0TsSI[AL q qbfOSinC={*pxIC)U%$al7U|xg`*8>e>`d>2ɛO=gm teBγNgMt LDCmكb!kR1ld60kR6aQDڜHxn"!nmDCrdm'}1ld> pmU3F`mj{Hue56mꝻµ\5X>΃U'>sP00$&X;2;E[orDf|#`bףkǀ ip428[p(xTRf\ /EWI( u֜ĪЮi,hjJnZ5̕ru2 .g$X :Pې'gR&gMxOKG֡uS'?59SV43>&g ꃈGI1rX(gdPmS"#;sD%X m+Ń |H02@h@鎈 `$;v[:=4=z' M4W bvsn*U-(L-ia'M :m:J26 )|n_O1GNB*<Β!ɭ62nC9ꏅ(V6BֈL8k8p7`8}NI42h>Chhhs6Cc 4vuF7D:κUYM֐@Ky:t(KOkv=oG[~ЩxpY ċ=@ ={.to$Adl ˣH"@dte_Y3w\OȰ@9̫7lEL#d!Xc 2!n*.Y?'(c xFC]p`G()62T㌪5TXČmFP5:[ej P8cQur!BLK& (Ȑ PT) 3(vё>@H0X=!+h %I02Px[[ߖ!bdCSY󩾾 e69mTJ t~9%mlfE }͠}DIXK+Ur:*gi"Sys ڳYcޯhY(`&U 6Wr2BݣŠfA`S&g[IQR%Ҋu*M)FR<{Ui-ԘhP1d%Xɨ4Y5*= 69S=ߠPI( 1o)'xZrŐJsR ;nS\EP+8TW;/8aS<* mFZ,)S*aJNFgJAFMC ]9ro ?P6X*s*9fB F*_Ck@xT޸J) 0ԑJ@:Hlڇ"qW.>VktB%'ʜ3^71fUp*WJZ|E 9g:[eN\(.ryk ~2mG`kpM)CqaV>Hkb7Iq%P1"l)6p>!@Bg" J3^k'n026M1]Bb?QSN;Ʈlʪv%SIZQ Ұg{!^ʊᵇF&ظ 6n_~1̄\2.I^ߑFgi2AQ Nn+8yRb%CQtJfH7g,uKR<>ԇGc{-󷸢˿{;[C<ιѱZipOmGJۂPp&TX9d3L4qMpJb E@)^P)Pۊ8\g#^T4G+H|{ϒQ" *7I$R2Y%\˩kao?b׵\A={IiQ0DR0&FjԘil9~?F NōK~ö%+kkT F+`I!26fT"$;ރ,104ߠY]rxS}Oaa~BɎo`a/M~lp෬#G$F- r-B ZӊЮH8 _FQfAv%@/f d,1[3Kw|ZJiIZ i&FXUYӊP(a9esaUq1N|5_|yGitNW*ࣙlUcO|f IlB&$E$1.g) 66)7!MHb܄$ΰ'jV*!/ +w0c)4flc6flcƾh3ZJf6c6$Wt>?N8@wA/3A N0JAZVu0:# arwn oI(aYm?œbBtqrR4f2b&2%a@G'4vF UqXX~2IӞv8ߍ`N^'3br1rO,S+*R+ LUfYc8X )-c1"zRtX`Es4hc~r|ݛW-BARmL*gA`OR[8EP'R%iTsIe#^]Mp6o)V`vv #ţLrvD ǃ@.8fǂR=^,Їu:*3N@n1<~XafFmgڂ6uu8vOބv-c1*%&(fJ ȗ3'6qg_tRh0}Q-AŽ|zC(ػoElI:3ϮZ`5A, B#x0]x34ʦЃ7:ǛMUkl8UET"aX+rhE0!Y\5=INfmx:};[uY B' Bׇb;-x+IT1bͫ"]jjGyD6">t4^KnQ/{s4eTZY4Le@,%~Eqq"9%d>cR="e4%nrρw3ɗْAR)V,xb$ķ g>J|YÁ\0*N/Fyaa\!m^c>8w%.Um4|;qttɸ{qcv-HH&%xAaߺ/Tj]L;|r=5>r^-rԤV=/!E㌱]GdʫO=0/3NӐ+,ˆkaIXvqxJV1:C̚Eh0 Ղ+}Ls=T |Ik?vr.(Jg P쬁:]+xsq~6颕w~,I?4;cN u?^vs&r*8  ̧{$~uOW:+O~ +}5} &#a;ϩw=) VD=AI4FFHW/naˇk`.>0wjI󂀴_ ̽[k@7?9Mj2x!}K ߍ{O 7Hܗ)l8fND6D**X,b a b8ZqAJõU*&VD " *K %) fD0v0vX=l Nr3{;T$aҥ(@RANa -|TRb,c̙K,0`yZFD1Js D!+K`hX s@7-SL46c!c"X֔d; zHҔsĄ r2_ɘ:4wP5$q7 KHHeaL l̔[74$M,:-)$d"m6H4af\צ]@q"A B8N kgi" H"8Naah~.o A b`A85I*ZJä> n ?Aʥ^6zM )0WKՌhVoDdی2"hO(\>"X!yQAq**uiSN`VByXqVLfj 詆&H4||&]T u9a f2aͷQ}% @ -!6nYQY'ys3 ukJ!΂"ؽ: | }c*dr=a {.ET j,~3XCfдζ+T\|FL%Rܼ͕{a㽎,+`H(@:* v i?mlB*vN%Dj^YHIFP뚋g2R {ʠ_N#Ԡ,E޵};cl $zuSH&_CU+c囯yk&ԙ: OA/'t[7`W$'cDSS;(/HD}  `Ⅳ~{]ׂ7-^wB-Ou"As^4I| x "ԁ#.XX~U)prt9|$\zU@sc Jo,r+,JQԒ"4D(2 # /C}2\+ g0hqT/>"$/u$`#fP܆`m@]KEUdA*X?Qy*NTetcE5Yrfe53Dv ,} O '`<;<|ڷ%}}6Fx4a=R߃&z:EjC$*rPK ||pD0-QLAQX1zYBm.FT (G:w,0 :xY+ lw,w3JDXj5Yj0ZZbA̓tz=EȒ,L\5vr`m LgAdD*PlV=jw7[15\-F?{a'TT(AW ;KB A169,`k._3v+u9MTU# r[7+Riފ{)pfZFk[A A@*?o7o{`?Rm0d+7Q;y<=&^D!8P!D0uq&M E=*588c˿τώ*'AŬ㮥bXc6YkA#󨑲ZthQi~$lH@{E} aPKdl )1lz(GZ}E:%\Uz  )c䐐A hoLYPXAշ߬׋afAT( #+Ij.UJӌ:Pjg 9ceXy*ažcO(eQF8UQ [od>?{]B*CSL.y zJ19I5,Q&L?X{z`jUZ^ڟT#ɐ+*R`9ZOה#kI#w ʷҌ/MP[`ւJ!#<VsVƣFXf*(-$NEzBjz~)"Q%^kOQғ옔C%I1)g\$!P](hIWr4b(*?t-+ Ay[V)z@nߝ^ܴ^_mn7hbT%$#L?v$);zR-US+~#:ifr_u{B}˻ l[sFiKM럮]t'9w5c?x>7ڸ qi~?ǫZ]O*'x/V=Vߑv B6XPYPõNGKtwC;*b?:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:C3:Cb:ߴ7 u"]PGKߓTҥ!= ubePguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPguPgu^P[KXZPp]Ѿ:@̋!= u<٧[>                       fBv3X;ߝիN~zwvo3&r-?_}\;56^M4Wxluz@˴7H`;@#?W߷}_f99]m^ 0xfnQT~ֶ>i\@m3vmǟnwg|KIO!~8sQі[kHRrJB"v.?vW6yk[]trq}Wnt4sW~_OG\Oί.q/'WW/wӟ[;o&xnmŪ+ȵ>%p7oq2kTT*HQޓdxe GOS*¥|KFx`cBz1/l r~ #;׊Rjaz3~RFVPxE{wṞbso%2vGbj&{8;XRF ,v+l;4vk / v xֽV4Bdֈ&v Kb̀cVXZ X+Dޡ;X׏<_Jz9`qvN <@\ꗓ![ޝ? oÛ+> p+a+>^76w5\QRͱw=v~N^Lm_on2/LgރV߶o>9лomGkvq\{P"^gO73?ݖou@kO]}<=]xշDx Zçn1qsBg)X˂?7WMvs VklM/b>4fo\|엿؃=/Mg.}ϝ>w~op7ovryWYşa3m<ѽ׷ Y[{MBQ_ ܢ||`g?|uV!/ǟͿf|b2~$>'gѻ1r? =B}9dߪJ޽QzVQܛ.ޘm`sM{oZњuV*fK[rNLq,'DG%zj/޴KrXJ4%"q1br!MԢ9Ul_ǘ;=?Mk;ͥMLС8cqܥ%}bSIkɽc@fajݷgv5 5^u`(j(5MR $΢z ggAl5\RDiӤJ`RRcUi=-=X3c[o! kl_H)J:r欪-u7@x¢5zpuNje]\݆Aٟvw!ٻ޸rW#`$Q$2#v| f1s-XR R{vzYVApXVŪsxyXƏSE4P&sA)]. 3p1Mwk(~u iCB$ ueԯ2im}yJ{xɊMqW*iü,srsX|S"B!5ysYUtP-w򦦒ZNŐj(%%銻#s2J <5vQ,QMq,ƤSEW'qu~0`ZS6^c{m~|.j() %!Bb N8i;s 5J`1בVLĺ` ɨ|S!Eb{6XM"c搃G)f-^wbP=5 Cm.5뎺aiR@x&P6İLVx,(<4nv,*60:-B \ 9Q3vaa`(g_e(e$kluXBs+}W 6l CK ґKqk +3\oSp. U,ev~c֣sV6+X)"2C5 E)H{/-ԓ)H]c]Ze R5!c4)*@F&22ȄWgw{mr@4(M餻`+,+urb2 V[d_Fo[`B]SN~j^Awq+~8B2n`Ɛ)u:rH0HY`.I\1謥uv&Cl(,XY0a( 0R8()ؙ#CJ/Q )PlAvNZC!'Hc9;Lผ}mo]jk2.di1 n9"ka&c[o,yi>@.` J&r+"*2a1=GM`]Έ j9J#HV 2ȇ`)|CH`)7zOxLC l(+d-"[@q;*q ?XW& :QjMU;yǨ碹",'K 4 XB;KR픶b:"J=tW;n%ӳC׀R"T2/Ex.ܖR%lKRJcEX0tt T+fَB8rFPd_-{:L?`ɐDdQuAb} \- PkBwMT~Bz+yG}@FѳW~a%^^ܼZ7[4ǩ>i8S"m9qzO~?JQ(=7~c{1<if`ĮJ`w*v*Lvtw%lw?OkX?17ϫ5>z}ykm‰Mi~-paVI{NVַWn`%Ol:Q>b:>u@/ݨZau@uFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFayFK2X킌:u_Qǒxu@F8:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:è3:èr:DK2pudu2o{]Qd NYFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFay1F;ߑ&_^t!;A~/Ͷx`p&Omi];nz_ΏNe d ~L;;>[,Y7n!`uB"?/,륄 ŀ A,,Rz) 2:(ظ#xX/fcm\Xk}K ي6p.,sq)1buP>n`5sڅV/i^Xk^Ⱦ}8ڧڒ!]7Fmk~]}8oGw;/τKgm_ Giљ E_9UDrn)Gٳ!:B;:>cx`ma]7ke yX;~#>6gZo~ৗmσwq]VMi&|_7M_)EV"Y|,I[ҭR*ccLF})]-ss-o Q{jrU3&!\(2xwGO//G;~Z^[7-Zy.b}t臫tWLz#RrIܜ^qU+\_yv&7Uբٺ_>G~uo$7PtrT]Ϳjjm=&rAN-_ɏyZ9=_Ogh8l:GlI>#^}u9vBJPs`];`٘jцоd|=;É"F-&i=6Z2f!`BJWŀ 2n!`WKuM;,jb{Z5,ed-Eu`3:wN`o6lszh{peyxEHڃn=pp^z\în= Ws4tF0Q+Y~ -G$oژ?SVu?ޯl:{n ?:G=5ϿwӚҔ/mGsg {y?'  b9OW׷uwPs|{Ig=O6lUZ]Dp ZR緧wtv|^]!Xn]nb )iG?xwTNglM̼1Is7єwzunpoF;̔}+BNm;JlmLU;].^Tvnab^ ܡ<_1,3eE͟7?sl><4y+nz?GSGl sIǽi϶&'mQHsu{N馔E|;=a7zJRQw_#597crS餜%=uEvލ5hrʤYUcC( uJU6g\H9WS5AUM=S{g[BZ9sC﵃ԭ/Rh]U.{&XC@x{yVU!kGiF6asRkPÊ3Z]܌6C׶?{׶FdbK[@c~1؇EcW[kIԈ=(ʺt%Mvv?ئDTVd9Yy2-Šcd!VZC湁.ARkD$k$$U@UgUib9Xb qo`D38f#nlNYVjdE8˯ቈF=ղeq%%wrvAHBM :JԸ-3K~ާ\q1HҚw-ql) ~%|m ~"gHnOjlc^/Ǻ:>dᝏYLW-guH,R$S2!۪'7EJqٸX ͭ$Ϫ,J̱6G$DE)QmDI2{G{@xj1Ii# atc3 !XGqu~1Jag!(XV.LJؔPR@"Q+QqIB*~6P(f+X'i.fnX(.z1F&l.-6JR`ƨ1K"[;ݮi˦PB@*!+ե$P2L5RP/~ F2ߞ rw ^b=^EDB.IH@/&9І(T`%\HZҍt ((_L(},Ci[SEk@Qa:ࡃ,K^BP!)vfM)Xj5Yҵd/x!8 xdI!.x]oPmBg$2S" ^(M6$byc%UjQPOJüO( AV8KF416ʉ6ؔOX+,2A5$RYLe#HRSi[y/VM*w+(a:H`@[i%W SL!*6hvA[-ns:Ma|L:[^ۜkYDҘ@E#!nnHh%GB† s0gBgG㪩bXcV5 UFP IO!H r7mc1Lm&di=QWTC(M@* 'Iyˈ*@P(Ų(Cc&UCQ,FB6YL ^EU,W (]ڈr1AT48vTvV댙 $AcMZ B&TT'Q !W(-(!Ts'kuNnY%kP|P|"$SBmEwUɚ1  &`<0Ҳ҂P/ZC iDF9LrƇJF9T5] =Gm(T q4 Vo^#pHP"l{l,1e\K}``eSFuL!jIR)gIBq{S \Pk\*?"t+yg! V/-r x.@5*Sb48O` @-gA^. Z"=Tp A'?ϛs#z\]n(fbGcOM9Nnv {~ћ?g]uZ>WR|ʟopi+r|xoy'eۿ~>>F&dqǨOj:@k^Q(QhvhOލ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨs,8es2^Qp]Q'pWo ؍:BF ٍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨӍ:ݨZ:ϩ#ZgcaDs10Z;(C7:PbtN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7tN7wԐS,|;A~ɸͶx`t.D=$ѯ٨}u1^^^,x pw6cl+LZRu$$KZw\tWVz&`!MedLȫ3kd\b֑3B'\&X  s' LJBfVY Xms m,Y% X]g#yF6,n~rOjp+:-. \2s fg5y=\ ;UVKt䜶;Xiǩ:DLjl,VL0L:~?I')e*{:NfV:- X4 ;pdn;XsKpLZGYh&`sI]y[` ;Ʌ X[vd6G3G^>Jʹye\z X0f&`7%)v\"cvkN#yX魔r&`wLj53˻xg X \R[=!p F;W,w+#fV(Z{γFH!^7*q7|^qi0A%ʳp=jKg06sI`ٙII3x ?ߺX{\~`,z7׵׼^,:YW'>i_>˓ѯƃ9^N[?MZbJ`ށ와\.σqst=Hgx Ȕݽw{z-jF;;tp}KZ/赧+F)}OWte1tA٤+]^RxQ:՟1]q/ݏg;VcqOon$G{L 'ܽ%g_e3"WߟL]?}=I)@1GQjo?˫ſ ѓ1Uo~7wGe}=Ʃ,Nk幵W 39`}_×T[#uW{滿|zv5] ^9><߿0O!߯^E8φ]߽^"Jᢼ)t׿>&<|jy^Ch}iS2eCax_>>\s9/&KVQ޶|9KWxXm8oOpcPgc%ZeI/|C,/rDm,ݝ wocj\D_:^uIyzFdsAVoԃf{{o7[gOit?w?&G244U`(Cpd콊Tt/b]SpKA*ܽCg{}wRzz|]hbZ`K9pl%+c>S-){ mR U6,(T>[ب+|7ߗ}(fZ+u`O˙}}MHk<=83^mVs):z$_AM7}C^:M?7v>M 0uZ3R)^[kƿ4b/$a|Q{a0O&;.arYtA/*? AC{^;MC܁/Lg~Yb-r볼{#O(Xēd4z巽Z=A{%,c~ƑX-Lorqaϴ'ncoIem-KQf̸EX,XdUJn‡~4>zd /;.JRe aĔw XZ?_>78z ~#OU4!1?k4O)j4ˏV]_W6Q7P-l* ڨ y "dꘘ7S U}C9~q7[Fx[85jgUE[<''ʏˇy"|/CenӢ]L>M,S|sssغ[ZAhbxsbhO?kW)n|Y]z]ձm]/Eox}8On~ZC׷܎~nxW6L>b׭4RSO Sw?#@N‡P*]klV;zsϟfݚC~Zm[ ԝ|yj}]aewк!Zo~\9o:o9QEǕl=ss?NRA[:3{p pj]Jxv} |I:i/:i=,>Ϳ1c+:hUeœkkF/4݇~bZ@"UX2yeePPY* RnzqL|Y曊KAjEOv2 C($8@N #/"bӷ|]Q<&OtN#[@ ?`qa˷(ҪǺ?\ŧ|1"lT $;OjȗjyF7dürl)vXiysaa׋0׋0׋0mģ?x /aypNzjh #x R¿W \\ُ; ]M/߫"DϋهR YKD 9ɋ]~Pnwhye%߯a)U)m'\pC8ap pTA:؉ã%ryO/+qLSFK®ݿ @R Ⱦ FLd*+2JFۚ")v!Ky0D=g$*;56Fw'ZLhq=EkhJፆ+(3; ^H8#[>v"hGO8TsY͸]R m5) ʛ\~pQ)Xdb+]K8rv/d6xUF%⍖#8MA)c'r q=ETxӨ(ͬ}rdˋ2t,t9GK8"熏&><ݟ9Im!8Esć$Sr˃th9G[Q=E.McU!rd, 2FQ Fm{j=rG#rQ=a.-꣉*^(v 7?%iq~/MP >lҪƏ)뺝{|<77'\~z-.3Z9{b6"}_/>[pOp7_Vx]n>8[J&R"<A=AUtXkБqߙ?LÑq{H9M],ł"Z,H C| s‰^D5vaka|E$fa a >&R:NDPie"F%Mʒ7Ѿf@EX}/[*X|^D6L(W2nơ};h"Aar<#*H[#1R(bunwq!yoP+9)/hp>|k+Y''_I,10ʓ2ۮ&m8b=* SL3v`98nsnB%ݩT U^ޕ83ZDVEKs`Rt&l|Dv> l` +[Iʕ.^MVxev!- ڤqONu A*yOab8+X/ Rd_q,P0Ȥ iے/|[] r݅q26Žg~ޑA5emyZ t`s`̴ =TPu%F:G8RYHUhK=5P}g`s8*)sHxWj Xg8^`L%"pnNꘅpGŔ?|Gʷ-bC/9enxq>/cʘwӪ93#R:pHJt&! 4k pش7Lؘ%=7]$ܛ'g^9ń,{Wm902D^g)y!*%Um{2!)r:}gjuA12]*seXpƔ_A4jCy_ͺw{*̐Vd[Dp;fTkbmj&I,bAu<=~zEw.pٻG7hHNnJA9 I ̓5=,yآƪ hzg<%'p0B78) {GMvy_bzuGܐ-([JXNvn_@32uv6몐%O\'|$j;ǔ4rzm 4OUBjb^E ݏm`Zby:dQI; r0f>-.ys$gtb\gܣBӦrb8# S,`U4LOG\ ^3ɍY9lk9QpW> 䵝3ncŽ> >ꨟ# u _Jzµ~u o{9JXk3 M Yl3KA#kiQ}:bL&h_.9JA9~kG!x?yOB~.souo~6{+ЎS-d4k& sp(XI. i~EWӬ%¥)h=s\"_ϨIXc)Jz{gVGz<2 Emev&Ƃa,u$"8V3 T+iif` Q}-Mʇ (,]ZT.]VI3Q" ud=ޘmAˮ|:uJReB]w1mۡey_JTz`H,3Ӳ<;-UK<*6JyreCҌ0!S^¾Ļ0ۛD]1Z(@h Q}ä%qy_nzߟs%2 $Ie 2Ifʋ ]:*b룉t x#("7Ӊء#1=/PCwV~BڿJ m/3KY0tB -er7JJGe[%pVisq, ;b=aԧ =QyJv|qt";$T '2! t? [fxYb/o(7NR<o3YO8ELQ#j8ucO(ڀ9=`;Y E:ؠYϊ"EN0x4ug\Dv1BŹ24]Fvg$ې y_Ak;TRg&kDaFo`F1BcfW Glyqm@EWޤZ^S lDŽԔn^1ZD$+`IÖj I:v:!L"+T>^-qB)[d- [NE5EL h}4"+t\_87itaKBwyA, |@@DŽ Uj@ü|\T9;s9 _-Tjx{8ncp!@8M`79.{p\@T3jג3}~ʶdSvK- 4fmJUzD:DgLȀDuQCtݜpPy^VfZ.%dË MX8E:rxk+2|\2#~]Qï½)ǧ}ZuX-^=>e-Gi8.T&=OER>cp8j3:CkR(uPRSW{cʝ V䴷64yy.[ZX A+0 L>EO ^\4bkEcljg=s-lG-vLeg6N`8%*Si8.1汽ÊQÊ~ RDK2^Q7 <|. of/t̀+*I"T4\|Jx*xPfzTz3}dvWF6#`\#,IL6vǀ%~_}[r$0" Q:s6Κ,Q= _)raRVLsIGè$RlL6BEQ&VUfqmg3r9OP,ǭ*hM5JDowyASV1 'Z^$I>Yغ"hTXQï/bbq"b2溬ǀΤLz R%,O` j] 6syx<4Qï9H9\ =V_|əKʠ ;2BzcDШ,Qï"e=\"#cRux@^0@Iqբ*wdlpK-5'Ou)e"ӥ0e;ENV)<@_’t.ܪ*B+93N娪4B`=t2CBZ$Ą4p|1&*D|;_>ȫ4sgA5Ft4rK›}~ИD E6v6˯wag}ȶ|4L6l~?, i d>v |E ,%]L$ko)|?C}?s~@w0VU~}[?Lݟq_7Iipp+.?nM>~iȒ`ӗ;'s>x7||MQ43bO fwP|M^헭Lo|'@b=>.5ۖ7 ,Jg~\MAٱ%1_$-_kw}sxV_M'`1^<ŧD9&}]֚).DpCLI6$3,1)پav8/dnQ~8Gn2IV*ƓqK+ʆħӢBeT|K䓖+Aϛ%(2}V?@E+5 *SL6Q)hMuCٔcS`ehgmi#|e1?J7jk[> <1B#[9Z\Q$3a!1SS(g{c_+amQ7tM}8wI;tWT^ܺf]J`jŠ6:pI= ׽<=[Np9SQu_jN ڋNUv<*Ч!tx>qoyuY!:Ku-^Sx4_|Ǟ_M;XphhdY% q狮ylDC"WRgIt^YHg5bDlGà!Nl F?Qt5ct0]zt.=CT<8?>?-ܜ* b `9TD[B<[F5ZwN`J<>l=o-애xB_ARؕkYl}t$>W \^U>KY+`3iF9Gjwy^˘׊DK1HqrCaPa5h9ls7t1"Spŏ`Z`TtW;px< iy@b_)racCx +!a*٫0__(|P6qqXK)oSJMBaX3)h`ZSxRQFeBK=kZ+P|wLC\?Dcaf!0:42,q6,nueIQ4d&$&b0_ώ [931 RRI'iB5WŠͭ"qAҘ$3pļHyd Z~&s-6VuLǓF3s;jjKQ5x~L. [ Q٨s_vqBI~M4VW'^! O*iѵTZ8șJ" 4IxC ꚋyxYsPcδش,}eJVԇ^ w-[1 GaRq!A-o_?+͠ d7G'n~5/k@>egGM7|[6cp ړ :sϏY; VHB%PRRd`Xx oxڠ$ʲTfR` ޗU3_,׫#oD;;$~+=^6ڡW>_oy#D7  {U\K@1g\MaLw^E (f'g%MQS(J;7Γss;vEi 夘@m0 El2phchߡk&Xa:iFL}po?*QC&G̵db3dޗmI 333JINk;LtfӲE ;iNеGd3 M'ㄐYH/Ʃ ~pphrpإ|MhYWkYu,NS^DVEQPaг/a ly\3ͬe;k bb.C܋eC%A T gYvK}&4<duPVuG. S;s{OVs lfy2*0ˉlLj) 祟ӕ?ԏ^Qm7e|}7>Gc|<&wcjᷣ|#֜Zׇ!!51F Etoh_Gn\ҋDk`+)BfE@JnL1^2 C)&kl98źǺkcxmacC\ jJӉu7w¬'.^Vl;e&fp'whY"Ҕ/ڬ+T;k m,g+2M^,v; 4Bv8n[ -nQp;|Smem ~K&CtpjʘG)^ zL;Cv~;,c7׹)%)&yXf)J| GN \"s}Va51F n9z`,O?>_CM9: &5+Xi#2 3C0!ʼ)3J{0|h -QmxSnڠ&뮕 ׌"SB5dS+BΘ"ɯ^/51F LR[{1 WL繛k9?FS`Ԝ9h2-G0Ǡϼb qwe$b`֋}m},5G&)a;Jd(A*.I`Fq~!)$֌lTi'ӊִK>LjY}!:s* 5;=$%+ 9ŗ{Ro|c!L-C`餹a$S (ўFgD"q R&#/V)]pWk}EGO$PNlz_3S1ANzC/"\^k+PV^e{h 5?ۃE ),!/cJӆ&'TMN΋sCcuRՖ$Ѡ 5&΋}7W,*~OhM?Yj)D`2s.)$ZñLd$*Ӟce]N'ͫEE̛>WDoo_q }s(wY[[]s N`+Ml`X%GTsozM]+e!Kĥ#! m'lm;Q٨c<:o,JzZOgxfܭPħM=tNI U" iaJ E3˟|_"C2 `FZg!׉D0ͪJ; r1}'<{++LA"J: )+$ q}/%dոSRH3V(bSײ1SgN;w?R($3&'5&JWNs6ewl"fP ww!`&vfj)y\3Kq\KxV/ &jG[{cZ\ox-}vrR^oj$o2i#Lry>,tc{0zL4>yw fvnC( OwF:C|bg|˾M׎>_OUZ\?*FQ-I3-PWD+ A{Juoc_W^%Po0ubD6n;ϕXdtqT}/ƪGYe}\lc+Oq&ٴ:> +n?Uopw{5N9 caic6)CJ+)3 3Oσ!3-KSgPR<3..~6PgsG)=6QmKuc ({4Of*LW U*sk 3n>e.A0a6$O'rT]0irN[zO?Cl3;싐3ʓ}:Wz@V`aȠe (-3`1&TQ97L]LNT)4K9vZ!njD1h(h7%[sC´Ҵ/z1`ym^m닞 L3d1k?O4X7NڿMN6tifolW"[,u򋗌q?=~ioq~;M"@, Rn࿒S6&?5ٜb<<=әI?I"`>*qԾS~s.gc˹WbC gpdj ßhӄ[ K)X>pBk7Y6@ΑLsmR?O&p^0/Xx08Mī`HJ!Y~Rr3yu2 &'utjâ>}~ڿ,hu$ݘ"bCP<1R֞y:[{g#z]3*K5N(6̦\sC❊zx&X.T WQg(x[&֨5X蝯C7vVXoϝ]=w'@Z%{>Ivx]K] r?鉊B7/хWsry0=[Q8wtxg~68v)*vۛ_IC]< uD1Uw'mUK]j?CHѝ3β<2 @I.' LIS`9?MIgDq! k>i6d_~a گ'l>egȶ0O' 6v2X`~٠xr1Jӳx0!ݛR^ R [yg xH3*OǮ>r5*jwI&k@J+ zJJ8uxCf1 JSu95c5QѐK#V:Wʰ# P*AZ?۠T#P1'ChvTSR{N&YZ3.CmsWw߮PYbEV6ʫ| Q% fQ$P+Kql'x~n'PR^i)N.Qa c]GWR^JUuRՖ$s(PRBvך e 8VKq!fEL|J\5.l"iAҀ0)2?nt!2Ú* ?ggZOAOq{uccRFT@%U]*X઺m 3jbdžB`(0N tFbd57<" fO0'v ̉"D*\v*.,~9_yNj= Y. cue>|p#|Q{/C _ܙ/Μ6 ߧ_61;́ےnޡK:-[R8.Wj %Ud헰{te_teu޾zEUJPz!?bM8uױR/MK hi`1`UrYK|-VEVr4=+GH3ّYݗdScaP*G&ūrou%p+I4y[SŴrg`+#(1 ɰ[7IGzm*IYSdm"Qg 7JZ 0M=tnc 2qQn]n1]-qNڑ~4OuI6s9uƸ*-\Q֜ԁgJ+F|-\ɣщ-|1_W!&l h S2bb(tp`DiEOy}ěH3b.Ҍ!>fc郝͜=>cQU!W}cP> )ث?vDB('&"g;ZKbڣPKrbl=5r?:qH`mP)9X1d0B!Tk΀9V!^O-%9'QuJxrMt[j$[G p\ujVs23vRzuL>\iV,FD_0v"EHG TD }%J >ӋSǠ~`ϺNqSQiVE* P)U;-**`5dpp|GȮ@Px}#x1 =Q+\0% lwDw豱ԔABZv >M )10ߺ&X`{/׼cLg0nxTnlALN]Z̦}2[.7f)EUto&,"Y½W~ӧ7qhJ33OWS$8 1Vqo{g_ήQŠ|8r88c~ ,?x:j2fmCrf9̭냹bR[ƍ/*Ϧxt1T=zo)YGz.Xh./@qv;fzw?m}=߮&qA-[*BAnQ@4#`c摰+Tx^fFLdzse;vX@_(βxXIdQL;@rZʥɇ ի^ى[}{@.#D Qv8f`ai8CSV%8vښ ά!EU[(∙H(Dr>b\~J,Lj^utHt6S(탼WsG5V_Ӊݥt)}/\_y4I ^eKK >{;#H^~xhtR4-*l>~_ nwEZ]Jנ3*m5#d4Yo_+d7ߜ$@NLn(/%JPƜcD+aĨBK4SQT7^ 7WS/>̝fndgyprr#v̰}E)('RÖkt2ĺ1<cWWGf)zfchT1mҡګ UER6z@ْmCy.mSf;%)$]ߖjNXэ#ɋ"3.߼r X8;^]xJQ_߱RF=0M p??,֬3ˁ/LjxA{Yg07&cODs^=NU긜0qiT)k6zUbҡj) WeE͇%I /p޾_]B&.!b.*F= U՟;w9$ZCˊw͟:t$ Ų֧Q+ټ$vGo Rhjf6J"50Goϣ.d].wTW!JzI6WW%'/1:]do`4b1G=zd˿W38Y$@.z!EZPi|X!!X* vs_R[  /qP}0/0Z -\kv-\81Iԙ8rY }&dpbW!q%BXR.hVʭ-`8h~Az>}`3"[Mڢ[ZH1J9D9Xq X!E!ge Cb,/>^zT#P\ۣ|xuo*W9)4~LcxZe- Jg&hZBw:(X$*Jb%f\#ZNaw~ΨnؒKI!s5OK b%miKV A 4֜G´Nx\k?TQ6ChF{#%sH ["w6K{|^<ė-D-klj(V6IB1 %*q\ 0pAl:o )n*${;آ|;כ0 2 Xx3U"5̸aD3s$c8k12ADZ"Ybl.ڐkA ZqXBWF 3S2ę[)a"aӵ J #,[˕j6\, Guyi;(Z {ǭokV_ iZXA?O>OX?|@P2c.B4,$pbB0M ݀9V!OcH cB+ D)D Gr SX y#`d3N%CfZR1hv*n䜵琢SgCx9ɃD>|i㆐JbDb\b1A:noz *E x>Qݭ6x6}+wGg9H.n8xrr6?u5cFƽw3CP8oiۍުʖ^7෕s:ca cDJE8Iw*iB`FCybt2aҮu ڜen OzMC>atx O#7z5aU.n4OQ7jI[&B4!uR _C%Zf/"]J}bI_{p_W טP*J.۲82wAF@ uv|)*u{߈C7SӾ S5pAb,-(< < +47Vq6a{EG]mW ?EZtMQ4hƲ]j#65gNj&jV<YH[x֒cj=sDJzVune"Iq ՀGy7o. g8kbЮi`)f=}{T2 j`7a !h-}8!uc 't |ⶌ8B }҈lӜ Hh 2GtzӢM>ZЌq'bفrڶ O х@p9Qh^ Tp|![Fp? wy'$xwlgD-sNRt+t`kՀXڃwDo\oZvYؤ1XͿd}e=pM׶eW&uSmCe`ĻM,׮>y Gضodȕ܃rDD6v]=l/4tQ"XC]mnd)>@PwGӭ85ˈ֯?S:j_S-G ݨg v&-4_{.8$hڳ&]Ȱ w&,9A^ +u O^c_= /qWnl{Hۀ*Ok0"cuћϫҎ XCщ]7׎'. 2s ,`~ǐȱUsℇ񜵝 <SEF A>#xz2m"ZNbgl@9!05O]C Xw<;x$c'{P{t`|te`݆A_PϘSEd QZi}jf"MQ8[9V #>C 7"ꜵt4O""@LO:zІQd:dp2$[/}hM5QD?N؉t܍[yPao{MÛ 4O')IB#C "e h D 1g\R1;2ĽF67D( ڏῙï7| "U Sr. n8!M[Bܢ! ;ml^)2 W Uap@'jdYI%C2#<.QUpN6 ,`[d Nޠq b͂s`Z(7aqZYguTd>>߬8l3[9lk1fMjz|)ƈh)ύmR"ُbgT:80Nx)2D Q)*į<7C&=Ew_D%LQ9V%4 18|kv~*UgOO30 hwO5{7d2=V>ZMae d" 9^ F,7%B "?0R<Z3n~+|o_JC\E@ q h #"nU|~s:+PKD `3ùGsL& Go #6¬#@!be090#PKn,{{.aO 3u#(hL2 _  !zɎ']!zMEHH*V& ٿ< J3t|Ƹ>7"9) ^1 y\rma"o9a)LzN F&C@p,&cGY@/nu>:kq$wrJF,vD94 18:+;`HJ@iQQ.b0NJemX^6VDF8<"уKDs?zv7#qu2kGQ0 舜V Eeah@FcpjDڛu{+o&(`o+sg>0[4H88UHzO*'0D[e]2pB2ڂ#,!r[d_  !:_a~k"߮WqբW7D׍Ju[Ј}~.fb*-yܶ8J6=E,ZOh[Y! < b02SЍNQ\ڨ$8D A(ї6AjQRQ4aP!wjP-Ɯ!x\̠V8jPO =EzcI4"Y^pjN2xʯPB3Hd"̈́xFZ$z=|[H4Lj*&OjkJLL!`n4"\{+r3\{ RG/n0s7`F-JY h4Eˁ% $R⣂Mc[.cNBO$RFD(+рTHoy{W94hǧ>*&ֱ}`  [ABLT;`G #eF&")p?  h NK{j[Jm!椈Meʈt2B9(9{1QJ˽0ROM\+OLm.W@æ`R##- f@aL[T~AN-ߩG}1??H0hM)̪Q.bH$>"-Ic/83^p*gad4G= Ԃ1(|P/㎷SnSV _7r)5/K* h5 9a47bbQЀ']b3T-x{H4^h &o)v8p'xpryWNDz2CGO҇׾PUs) 3}_%+*w91 "5E_t:֢uzR=S }W#"Ay8d|^DRmRp^v2ڂCiuZDtz)jv|Ƹ*Lj s]Oʣ+F#X8"e҄9'ߤ)*;Q_=^1@ ) ☛RPDNs/IB5Sf{ ڡjDν$9 18 z|zAak'#$8 iRX &4ڻ)CɩoQ˳ɨfi&Q^1QD JQGiL/;h}X.V%'."1B5.?դ I4G\R3+ ir} JEs-,_8f*<_Ǘ㞆ցPո{** Ld$I1wQ\6 18Y>sc\%sz\p:Srɛ7BSwf%*vy+1H^F+22S]?_e2R @N(C30Õx퀌.O®ɮ||Nݲaϱ٩+@Y&딏 騢sɿL]W= 6kC׷jǠ~!,=&`]S^ WĞ"2G3-9 >e=p?tI0 M=W㥫~~|Ҧj_+Q!>F݈2 ms_gΐQ +#<!Nf!_k+ЀTH)U[5r Yq>ń2/ǛSl@Fcp*un h /b}K"2S!}Ҹez潔u{ﺧfiK˷#&TD踈BVc_m@Fcp*+~ֆа|e@xn~ڜrfYD)DCXQ`4"qQpűu@ S&^k6ꐐQ]+ٍhQz@F[pTnrSN,_ 'JHCᨍ1hpvل&D h dXQɉk4q_#Ts.r1X3*+$&D>ዺv%O*_y\۫JG+=)}-0[xx/(V"d_m^Œ LYK pAC:dRDP&)s^"!gY/ukNL[1o +(YLD .P (恼exE>/Eb;~{9"DRt `@7%#G&B>\M|M,9TD2VobZE2M&,G>i,}&Fw]_Q M1}\[xn{?x{o~OlR%/j6ΛAViü+my=h߿y=hd?E\y]L̈9̡NT7u|nNx{LOvE`anzm6b*Ρ_fn;׿ VK@ 쑄 :w3ۀjs\f?>)z70O^og;,YlVw `?8)R|mo<߅`5>6 ~ _VL=\ט;>2z%,nP ?$8Ľt{-{%<fB& bRp_9GNd+z*F$Q(fO3lbw?f[a2{J z7yngX(/2gu<0gp8̵YzbO^ LXlߺK7^`-gu/0 //)Wt^(CA%>0v({#S{n{ﬣq:СA0&מ&25ԔF)u< fޕ6B`SRއg`Oe=^!n٤LRn7xD() hfq|Hm@ѴFi3gNǑ#.@C2Bi-Պj\1. &o0|jj\, ܰ )ΕIDGApM$GRvJ|Kӎ.+nN̗npinpi覺4Tr<;b!h1R"E ^'][:vH:v~th@dc0r #70r #7ՅtPFj,Bt kI$Jo-YX:1흩-M;j$M;?Ej쓸{$$Y??F\}fy h/"EqWB #!vz?-8,7?74 I2QL֓'G V KHsDEDx6 ^ ̔ipcYqŸ؈qYɰ=fj%:Lm 1U mK YHuTwQMuaip0E9]x̴:m0) هi8RZ!8d6wSW!>iy,L( 0h$X -%{0M;jCv@V8d⊚ͳff ̚5+Y:j4H$秓*Xټ=7glTlVL.eX!)[ς2F̀ތbJi (QʲGu0sεլiGԬiesZD59e@$sYjUOt[|5ˀK६Tx)s¼ڏ\_\G~{RiO\=̕աSO sG1w*_,OA5;< bV3܇5 Y!3uL)G /L%{)dzfK_"Ԥo{M])Y;5s)+U;u]Q:Z$*r!I2zZH^^i񓪞vCs*{UޛMQǚPk_U~s㓹b}K6 "tw׽ǃ]C򋹻ПNɐć uECY".@HxTſgX\KrPWw%),eFy4STtrXph]\ p% jzs  o/gS3rZQmmKY纺.4B'$BWlMw(ilaKW|KBW/Xp-Iަq䕼uWy6#׉ڒz kB-02\'uñaRj}%xɩOJ_g{F{qܿ}I].ċ\FAeF'#سN"zI$WsN"Jzv겓o'; ̴){rH ]KX!8RdCJNSTL-J(,*v+:}nMľ0Z_4;\s9qJhI9|s.ņ> - сV#MLK$ޖRZn<M. ;*VZ5>}mojՁ \l8ݺ1w*hImBj,ڡ{{[]lGSކfQߦfQAOM᧖:M@f0..5{IíC]ɻ‹,CN,ë5|B ޾)#XãeD[I[bZdT4kgzӴ)[X:"+Pg\NԹ+S7WH%ӝzJ \Br䮐Z} 3Wo\aY宀`٘+$Ws1WH9Ri3o\YBӅP9c0J"BU睍m2V jiFj2i50UGx7,[HɉGX 6)KA+^@#*'BD&΅NFrC S"Hf۵,Čp42̈9Rd>$NKlEAReATuՓ6B#1It{Mf,)'ӌoEIKXҸECOuDg D*b)Sʼn І##(iڟCIc֋S->UYjF=_F>5f Q7⥴MrQѿV/Q迢R3(sͼ^f0n"хKJn),.s"Ԓ^BXm|,EKsFD ޽ЈZX,BɥTZgz)9 !J1fY*(DMǬ %wn*.-}.F7Mwܚ1CFy\_~φ_OSh=|wwCK1I3_1G/DزßFӹɬ?tRIগ=l 8~Qx퀞ͬWهD\ۖY:캉aǛuki,)%0p[O*{ D,4\?7<``$1_mq/.Irk1qc( DT&"-OD٨hBQ_$?,Y DӃOŇg $T>\]Oû`tMJ3e ڧ`ޗiV O}q^DG{ L. waaoOїauThuf ٲbWz&e }l2S^w)@fC=4wb fUˊl?e#Wn|5+.NV0 ]c0)Rts8_>N*:,RX+hS!>MH.粩'7P[) vzNCZwQu!k a Dz ZҚ I.(4qo t:g*EKZy]PNPT^ qݕ tߗ^W0'S[ֺ5L_$25V:;CSWD`#e{ct8k>u}%9-~;#0M|=} (G7j!HScr7傩LKDM$*\&v#}arVprcRW+x<{oW<ƥ% W.{`Yǡǫ!nv4;[yO\[GaZvIL]Rjգ,<E*hb%m﷔Zmc!(-qEYY$yoh_ֺ/_L0ގw堅t+r֝#X{y#x'q6C*s/]>O=\LeTlY7ޜ<{k9흂*fCk,O%6||yhEwDAdK޶ :=i𢘌§4)YxgR_Nd>?ד4NL c& L}T׷P4w'J[φm>x#]#H#WͿ-Z;ob=坷[b&311 tmKG wiU[ RFS V &cEY8C ˠkqveoazEl6dg/^NM>̟}P`eПL;Pnx@!G3.p^:Z$f&lV2 ڴaVt^iXJ`*8S*4(SL$( |-(!

fIdsmhؓKc"f)ZF$<2TS ې4ӴNӎigaGPgK3:i_Ug7?^6"F$C%ޕqd/#iCq g[6S"Uw(m)R#wxU` E{i_ >K?;(RdKJo+,'|&E tt1ScB)7j<[֑aYqGkZ ј3|N*lT햮Mܾ`qhuj4rdIMań6gW%H˺u)[eYl]n7-whB077{=KC-h=l:ұ7z8XPuM N6ҜMtsѬn{Nϧ~b70}-X)jiC HeԷ_Z9/d؅RjJCg(C³p<[]J,,ڀU'ec؀:N%m]!ȠHcT%&7 hMs  t8>08҄T_}ڠ[W .fޖF\2kT aeΠF|a&f,j>\7U|=OcYe)UC* BkK4*5AR鄡ڻ/[7UJmif4҄Ƈ.Le jzc`) Œ9K˭.#KXTK_NqXtW f.s*k>u{mP cR+mڌZBx>+M{5<(e˚N2رUrg-PxFUɳ5Jx?LIƽ0)Fi*/hfU(VE(Ap=)_1:g q1bw58h:ZɫmꞮ8%σ"]*c0h%g߮x0j|{:8])ueǀ=E±ŶIW#B>eKg)F +)$:;Mz![RPԧ0sz.܇h4PCurN9PNaNɶykۜ yyϳ}2KhI9i% [)2DzS,q>?%ёMY9p &W,OUTh)O>x%f> Bܜy'.=o>^+Ձs X2E"@# l-gkZ(#' ؂8cv`-ki"viu2aR/1 ,DDꥦ1+Cʘtrgiy=ƥ/_ hܻmpd˦"U+tbA_.l=1~t[iKOf L"ci9G-R``oJq&N9ịU :D)݂6ʌ ;6`#*҄HSELpsJXB;9{w6ݥ)vE>;AD>,qM.|T]ik:0*`IFTRJm@xX;zG٨|qo+ i[N!&1T)b+t@HmOjRz\Tkni9&nϘdwrCI{?>T9}-z߶-(qHHHKtvJOj@[XT ʔ8SƶJX}^4_=̻OYE* {zǀzcĺzB2<=@tQj"%7pFЌR#s!g|YlK>Oexuk]f_j@XoIkƠwyy%"xwjVwD\ \g5٠uFYpsWyΎ #Dڕ^T)[0 ,M'7XH>2E4^JJ8%\b%0l\'M ;y]s[^֜$ւ۳QJV|`äk3s?/Y|Y̙+833ff1e9t _]2S<2S ex@nP;In7Ԯasn,{rAvmqFMJ,垒[:ϒn-P7 czŰxŽ%?Lg1HPCj2YV% ZĘwm|;mzL^)LI=זef]|3t uD$]Wr3{WO :P$8y;:8l]+IBFR*e%GG㠥ZhMf?9p9&|ތ X\+V? m܋/@ z8>pU!k<;M/1lzIZ Cj,W{B &7 mO8b{/@WL9Ka]@ͧCʟT݇AivR?A_ 9v֖R2|%gS4ak!YN&Û % Ƙct_jwΛ7۴5i2bHkpZ:sbhS]-Y27^]D6'N'c~-Vl4#3N ,ժrtw+\v mRHAGC;>[ ]A`᫩ \U0(JQg֫_ɴQں6]T850x?LE(ƽ0)F"&Iu/ XY $UŇ_ׯs:CT|3:X? ,~II@7*vVz~,be!APvC rgw4=W>mH(T+SZ| SM}IW+٪#H,ۧD IIU{op_Ǔ~u?]gr0HSi ƥFXTRD9Ps<.E-7p]lqffiEJ3̏ G_^"o~9 *Q2^Eːu B*Caj3jX$"﵌FMFSY?iv˸ʷI#MxEJdXPd=L m a$*rVݦfHjbL R!8>jK ^Ȍ ^9Ǚ ?\]1ҟޤk<ꗕ,2`}$[_ude@%7 j-$* %4HHLBDN.d@ֲψ"epGf9AQN&=rɃ4 {p y,310BB23uS/N Y$53jEtck3eN˜vHN;i;Sl C+ժ:]˳osρG `c~2k.h}/EuWPmr5:Hɞ+amDg8*¾6<1C@*NT6bv("X K*Y0R`D$4Jr-ǎkAJ-^J#K\D2YGStd8ݴ<2f+% 4fsE5~s ujoJ)i6ƺ`Z:P ț "%u-aN 3v¼ z{JGf6Uy5>6M/uًns{ |#7v$o--o5LF$ n8NEMmNR*"HpV;>{Oa&?DX<)\_!%qoZ<ݏ{bzΎ:%68Mkp?+{6~h)xPƒn@`[4xCMm#4x?"ϗ %;Yh4Zx_?~uuWԝ_6oU>0ۨbm7wMjY:Vzy!j,o&Cxcy WiJt2*.(h6#DmO]\.d]{ F+|*,^RZ;3Ebb̍~g;o)GP-j~.p$oK򅬔C[\C֘" cM\:hC1_l!gJE:AO]}Wa|t&al"bKuȴU®Ts#)di7qQ[*nq]^Oj|e݃dwط9X /fdrA@ j [;-J|fx9sӰݐ6zK٘4&P}HjL?~P0O.p|{s(]O^@-6ePAl)p0VDM6[&ߢ.&|uKZzkab@*iQm{bzK4D1dJY(AGȾE]Lw>y b@,yWIpdMفȖK<}ZmYvt,2$ԭ[ _Ҡ%zF+[8p:=򜴘`R10B#4,tt9K;8tsuT,@z-șWih|,;[m)Z9s>R)CXx˒JCFZiHrym{A6E;XepF '/w>"13DTu"c$ɉʪzwZ8ἒ.u1/KVx07Ei-v>EBluLlO;^vN DDžM⩽BI$`5װ1 opMµ5Y*$% opyF=-"øކ={S@\8ǚO@/4DTVI*!vEqWN 'OOH))P[,YҖY/rN 'ʧGaw?cDf.RwqGL}c~u '/ )FZ3Ϛ3dɈL12n59T H7Z8||: nhAQOJ@\uN Grݏo7~C',+Uͧ:'^> Zk̥E'h³,\pxη}zUUi1?K_*Z mQv?ۿ-OlQT7̭oQECiNSsj1 >ϼfPxͩ;- 秜"g$%%VSr zU5$ZZ1$+qWx!pH|ycaA#b6+;-q3:i6Gv`q[WhI/д7hmH04!hWР<{;-p^Vz )L_{ﯰywZ8|nr۔a̚J)$3:\8Σrvވn[pvIj;-p^c;Mr,qsh*ARnBc2]97Ӌ9G {h7M^ZiR9|G d5_H"hװ{Z;|Z竜3&< Jc wZ8P=F7ci~0̲GN G7YJk(\X͘/HΝN83hQ,5Y$Y Kr٫L*p9\unĮڐP9AEi _:W蟏g&Vz'.u0}+?;x ?G~y?{:Cɫ8DsouxO%RK>y]\Obކ_go7YSl k?g3/\ Doxڧ }\7ؿ#=<š/xaX7w>/w[o3p_5r_g5;|.+^exSq{[EK)f݋cқD}qO'EwQ2 t[ls_?9~aoޡ5h~>70PoކW+qƻ&KwPASU3$[$5p}̺ [.d)MJqw'ˮ4/cXu6df|goTFr/DhTҮ֚`4Yl wOݭ%kcyh-VSkI~m 7~w~X;1{LN̸ɜSʫJˋn~6׍X2ad0 #1woRq ZOZw_aEDs=c}z.z$>Ȼ4QMq":(Z7 C;._P]Yͽv5~2AH*q~!`ׯ0ɷ7jzk.VqO()(@-](+xk.V)>n>4=WZcLv͙sPJhNMiۍE5 WN>MOh7zZYcl+jyre?Y0 {=WRRa$:LR/҆t=\13xFS|h]iB@J$+@|Lm;2,1{ |h,,Q n!w evި #Y}zhW0:@Zzdf)M'Q[2c`QmZ\ yljە3Qz4 LTL,K/ӻ_y:- ȭM_EXElT9o?P,:DС(=mM0(&J# e㑱hش4v[Uy&J ,@,[d2Ml2  JQ|R4Ļ5@j:jLy j3t'G3L mV@ !46HD@ jL+me<" c{BDV nBr `u!K s-,ՠ@鱱Mᔳl2beM};TtgD`(q &E UPx7D)$hFc\(5JG)!te?QR2}^cmpJȼ7oqΕW=)bF@BxX-@M7Ju{%آ@j>&ݺVR:L-50o}ɺgk?bG\Ҥ3f[sh!mUNN~fBl@ߗ {:^30^z~|'?j`2_sIw vUڀZ$ >:|*:pc*-鳒Itj 6866W/9j,$tG\"kvHt*o2ȇDR1Fgq{ċW URR4Qx(wm@¯^2eC|ӂ>WC"2-8]Ps .[C$ؙb|E{1w:9Y cG+xE4'ށ]rc8UЋ:"h ZU[׍Fe`/V ?xΣU3(Zi_!:DX(tP қUP!yv+Nڴ壜1 @ƈzIAuw3F B8 x!N>ǞccQm"B(HLA>QhAQbv)PD⧽֌PQ0Λ ;#κ&'91L22Z{0^0}hϢ;s ()2loP*sKoyP5ƂZ6A;*w(6JP_{ے&*!/ f؊qHܞyh ^lew|>bnqebٯ +xc@njas014mQf1:/tkqnAq>`_z GiFd9h7$%<%*jerZJahaBˢ`d$CGiCv-HU=i,K?=1M6BRésW7N41S u4wqɪlA&XP rGRnlI:vJQ[)S!=v^ D?64r,iBZd{  }]O;8%ƇJF4>=#z-bz9_U/ɁĈ$FWXIJTI"$ ET^4{jzꮯ \OaFU^5 % CKJ/< kJ:g}! UfԺ*"(eC1PfVu>"4$ ~K(A[@$X֘0u+fw2Ym"TP0jz@e d 8 s 'XFd`1"*@ @dZ|&/E!@2~#H!iUy)&6j3?jA;hTm 7po K5=c<;]DK({2DC Z_^jˑC̚pU(xC>V^޹T%XGI`^Ûԍj.CٗZxy-,][ޝ.擐]bt Tti>{&ՖZd}D&Sd!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2>,Hótj0t Tܙ@UJc @><@d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2/\`6S+:SJ3@J- @B% L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@z?L A 5o݄Y@fxw+H &Z%(P #WDΞTmBG .}z@:dΘ*.W]1WUZyJ,hdu\sN;cv\Ui%?wsUT4WN!u:|M/\/׿,Ao|0, =cAS"J\uĈXC4LT"Rb &$8 41"OlB=b_{<xк!RApx8Sz ޼>LOS}LƘG7.|~wT~O=]'.zH_3a> ȒN?@f~^PЋdDyOJ"I!*2fa~jbpmV:]:&RUys,'UϯQ&!|^toA`Tc(0*"GcdY FZXES=-{*!\LK=ª)9oR*(ziTyaLN(+i arlAR;:R%I\Vy!1Ch2X.I<֐IO=E7\dȅ[E1dkr"dLjUؒZm7 2r 뉑!R9!gZt읁swF\3zȽQ6X;֮3OgGSm? 14}+ᢸ7/#^zСNy ;\ƛ;/aȏ9sږ L_[ȷsr?)e4'#V1-8-^zdՕv:iKmYjh|2eA1meRa}$ۅMOU]'΋]fư~= X]xh6w|gx`Ņُewc*ð>{y c?"z3O=}Krm@qE+s7h<}ҜlKp@̾]q{?o_}YdtcZݎ%sӻôM>`^8jγv;Mk[sQ«]3vL|T3etpk|3Cȹrv<`f_|Mn.j7_$c[V$ ^2.'y?pYCm`y< 4Lf4#p_䥃^/l`5 T9rtqkom~m~| ?Tc\hdI'Hrx J~ 6?ygȭn`YvhJ0nAomGZö[rߠK>=Rh[@o5?m:Eh(#RB*Xe <:G}3GTk89Ik؆2qIx]T5kIiâtڔK:>b)[GlM 8 {O' GH}=x6z7767:ϊg잮g]&M0j[tv`!w6yR#w]^?'bZhݼmMww9nݣt- $zxIϋ;>Ra4o~w{s=pPOup@G޺:Mwwp(7w mjtVD|78Y/{n蛟7wR qC]a?_ _o~iF:HYT\PeMN2H *8RxM"y0٢ >eNдJ]8YKQ+4HRzDz.ɔAKa݋F2sր'5<::~dSqb]7cko??UEzo.P^„өe6lWxzmB=w s(ԙY>sfonӵmT0QHT:pUZ"c<9cuJ hNQ^2}dשZ^8s&2b.11M@[`_.ϳ`R_`{q|pg3:OxlO+!`N~wup:Ӄ߫ GmT#!lv](Gg.oXliiUV;>gvQ64Eo]$J]OJhԲ0ƲHI z LIrMh5hkiE{o7;BZje,rAe$dek*RKj -6O+ҶO}4sbqr=}QM<4) [#(+G?4}-'b]I@hh\mJFZG()K5 {J& 8Jؒ| \ Qeeg&0l=+3ЫجRݶ;@&Jpjmj`%sPd>5q#;I6x<#0yZGƤZZ>i|֘f?ޥp1bLBҒizS- \FAmEygWf6py/7~lߜȿQT)zXSz45'&pAbiW2ʉJ0K ,h[ {d]A7W۳կ0oHB8mIQU ZO])oF| l7jDDޓuu' 3[Wh"*c9lgP3hSY5Vc<&#uyow.TZZmD_/E_QH !Ypc(j'Sޖ6NZzy {AoK3,ax2 O(l)XQN;|*pM/7o <{%g۔(MJ6*)'cwM©-Z'VuN5OR^#3D [ĶOӘoo|)aLƣI#xEd؜0Ņ"(ǒ{8[ni^9̭7jbE;â&3cϗ+Ty5xEʌoNof[apFPjLL_4I1]Nfp,4$&La,UYfM'q4(Eq1bo){Y2.V2T PNז8U4_muuՏjon6O~qݰ͔`} sr,={lS{]@j(#4yi,D4DPsJuArib(ExteZFmMI@kfmm*ۂFL*.sHorҠ'* -ck췌QIlak3㥶жl -£G<:Lk-ӛͩq*BW&Z-.qZWtY:]b.4ۆOi {DmhXQ` `\(eJVGP(!\!h*а'ۗccHǜej#Zg%sE&TE )D+jfD18*a2e:@5 5 5혚i ;R "7&jWՉL5M pӄÀM3T xx?(RFI&~eW,@ D 3Jh-*;.=FǼڴӹjmE,V[}vma'җ~nYJRR)PXՙU\R{JRZ>`e+-jev\Ui?wsRj#+c]BWU`c:cNt\=R +9uC fu\UqmgH٣*Bt\)2û\?5yFψ(5KXR2mcd 5VdQ h@=ET0HF9[|+ _]tԙ6Uu1ɑsbAs  B kiyMm oӬu!N4m5F>f<"3ĄʙeDA阬4 g Pͱ:V Xu`%C4Ui X!T^WMǔCU%Fz@ZrS Xɜ4}"A&Jx-\ZWu2m˨ZX[m+CݕK & ѡS(", tLBрLIa-Bۺ=Vr *e9+&쬴I~pS_j^!B>Bmw ʐmCys52D0zݭF/W pGҨ}u{u֬٩\? _ ]F84+Zu.A?=Ҿsh~Ԑ 7œe`xp<: }rsM^DG_[EL<\Pz̡;]VP3sveP~ܩLE ;Tm@)ϜYjYFfdŋR\Q[\e-<9#3@sچ;7FΖcx=f.8!a zA?n`eɤOz} =m}݌jvb>8" <( RnQHm63&ChTjPO<`ޠP2u%$+3f&pCm$(Pra ng8Cv89V?fnw&ʧݬsO1޹گhNJ3j [Ô{)=ce>t7K~滏z_|.GwmgHVWs&"e/FBϵjٚ4,Ak}v.cjFXͱx {05۫Z<@t6)GMe,1@Fe dA&Ȥ[ښiws~e2@8CA$d0pAʬ5JiWp`ݖ9929 #帡: ^X<ה\ƕ飨߇6o5A7"׍ƱEGGBG`уr(4w8JY@J U bkW2GGdYڪd貏Eiڛ)2Ics\3BGL]4fFK "MK ƶ?AgR g08juP~ i9d:'zK1ԯaL9g,4O|*6zeE 9X7C·V4']eC)Vx z{ *šM4\-gK6d(߶͑ͩi~,r;+o2& tѵ$ʂQfɶWZO_1+tFW^֭+9qՑڵ^z쏁a$cl!~3Zi$`1#modQ@wVksʟ*af3'.x|ho3 ׬ϒ'rse[\"`y<\[ ]!WWGV(Y(CL<准iKok0;Ŕn"I&i ?Ȇm:Ƿ*UV˝[nN^p u/(I\C>V oL|oqhbNwՆ{>po:eݞ A LR}5 4j3{Zon^KVJ5fSշ(羬yOXɼ,T3օ&IvE_iglg.imPT8Nx 4 ں펵(Rse^=i L1E=`*YU[\QrjIg/X:yUinϤjoYg].3%\V4\iSbTQ&#J 3Ͷ-VyP( If&KVぅFs!1kMMhu_Slt=;;,vՕvh[O%:Jtp;ᗻAw*Ϸ>yrT@_yˍe5ePϟKVZyIU"YGgQ)9p'"p4=p,*󤽰F52:ǘF:4&feQ`:fyt)J$\F%մfl͚q=J9.l3vՅa]hZ]xYjci6b{i|Nrm |q^݄2Fd@ k"{.xhB+ۈ8 jl`XC4hӄOl9k/ 5!aYg B] Fa #O(84sMHأe+c Q4 NYi& ģiL 4!c+6ZIk%CJYKbC˸VQN8AVx欈6{8{sVÐ  ]iiK' FbN&1/d6혗R?P:a H;}>qf l SB#A* R! _nkU1iUa3h0|߯oLӝk E@QRS %2#!Mbp43R߼w~Bt[˧|E!غ!uuTmW5Y)]"qWkI,XJ,I &wUgϡ^98ȥo5+uuv4rQ>P'ߺrI.Xq<R]!sG;]Z>#p};-6uV u \=w2/G">~ּ`>tkz^׋W^rpYpd] R_uiP* Tu (E93]W$H ?ܾ cp))ic:I&\ 48Y:P|7.x0.InAT1zck5&rv<"|ZNWќF:W^wSCgK'c)o)ȯw3T U>LSA1TɸL,$$L3nAaؕ JFI 2@-[e r!4q+3ҒSPT΀QW?}/@Ս6|ɺkOaln$u0`;%#?,(nh k;YwF̗P8EmmT.Tlŋ bR:{gșon›Ovy=߻WsP os0(v|)\Z*7c9crs[7: Oto[njs+r/ˢ[npAw]gv.Սj  F/dM{B0TEXH$"fujlێ'2t;TE.JYAjBtO*V6!h=7K):Jm1f̖ӬٞBJƅ.RYJ6S )8 3k {?"7S1r6'&4rkw[,:o8:h{;3r4dBI"[vHR*S+@d n_WyOLTx,a!FF\d ؎!Et߆Ds9xgSdk;J".9 D5yD e$Si)H##zƾQLB.f4 иp ;&Vi蟳-0Jn-[ad LiJiR\7Ĭ~D+򍷩)sV|wtnMQ4e82g $CuQ2ȜL F&Ev!$: D YIY*tZZmF9*X -4JDr(I;AmSOcPe2'sՎ&º U9xb\D*#{nP'13Ck4m%HQjRƱd6۔I9-MmB1pz)ζfi]SAC([ $c 5aZ"~Et{OF_!%YT!$^3،cdYèSL6)xV^t͙!!8dܥ_}xㅩsJEz0{iKk70#uKNrnI>Mkbݽ?h%Sdy8"jٗ)b#9SrcU!ߢ+9)ܔer O$à c՟k23Hsk2W#m~Gu3 6qN!?.LI@)HXW29"tɣ{9+}X,^ #%¯T T.AUo}տZdAv؛%&H#nRPA !X Q< bj}])dctV]dvf,sH GRP_s"K5 " @,b0&,;_j+ԓ^NTkN)/~W L/rV ,,: imw1ٙ.f)"M̴DwOf=um!TEk?zam!miKR 짴*>-I[?7ʂpMƀ ggU4{x)D؊3XM|~2ǿ-QRcY )W o e_xxD,xBuc+^mt[4Rͤ(UL멸l]գo`U_o>ڍjҎd4(+{ {c2i;50s!{h jAKSg|I\u]ո6T`*U0xa2M]7VsW;]>}+;@DXdLH@)Aq""G)QT )*(& SՋ˃YeGg壀Q6q='3'eIu]YMqlMjT a *4"^$$`R/ Cxd!RM)#"b>FQ4`pHnI8[f7qoGvt}Wxypq=t9_rf՘Htݩ\=XL`D"Q14Z!`igvț@U vh4NyQ@QAsglrDTGP5@T!2f ^g8%8U^,}ubUj,fYj&:~pMy*RԜcMӘ_U;`s`a7* VB|E4i\T=%l.jyu]k]*r&D+1m&brq}4eן)iuS3Nf֗wܼvs*(J' 9v4ǴU=N)bxЄ^r+r.`9VG8w<|sGX;zc3SˈsM 2@Wv(Z]>swvq.O y@ԙ6m Ts v|Y+ kjífl*lR =twkߴB{Êxs7ۋ}7^ 4S7Sts-F~-R!rFg`C KA :Tq[`j- qRทREo{*q

t距}s,([wAA[oBRK`6s=2*oڨd&2\pcPgpB3J= ]S5C\\>DNihN?/ë$2X3=z~.N4{mL5:a mI\,~|^O@]Pr]'_~ȹ>uNoVP;C_Jg6޷x3:$`.a"\N=}ޢ*Mz^"-jR92e.̎0;<a.X#VGL3PGp`ZJ6NzHm[mlodKҶ hx xQC%;.wJ\nqet[wY\iqؗ8 c hy`r) ?Z|_q'VgarÜn -$7.ף5Tڌ>&1~;gLAAe&S$7k>Ԉy'NҜN@t׮|_w^lSs0-nG-9NZ~ty"׃S$$`-X,XψqTyq8Z&D-lV |On)Xu? n<G27F8> +[OhQm:y^ue,8t8NEw6Sԙ%a{g?7]pY_8nΪ!oÓL$V?~? C /yQ\ǿ\ټ@A "C&Wz ׵f!=G']VgoHfxcAR~m0»eZ}zl%߫Mn˛&Hju{_=ʿ6j͇@sP^03e|/wkBN|z0vڢ3NU=p2R/R=cg8f҇?r~vyUS3#-aO1*^`isB+5ܘ+r=pN=K+㉩ʔѵe{O($C0!B*Caj3jX$5e4zl5ͭ`Jv_}z`\@XJ瑌& lDh9z y:";ZS\vIAs>*yT OB A8GWK۪/ (WKM)qZFe)A@X/5^#ZeΖSS} TeQB9acrq6'_4޸t"LQߛ֒D ȦG'$:sP(` 8־Tۧ̒q8+Wʏߪgds uY3T.[ai<./Ql0dQMKNoz<0*R_K1fUY{gp,q"ך"Peځ@-Di2`h|RO*Uk4S40˛-F/]=upuTmV=ȑc>,6rCVC n#wGx`(z|Q }<ˀ{Z 9 ⺅odX g+HŸNz+\x.))>f^JDoTms 4 aH!7& xZĜSMExzy&@P0u=$A(( {3l7.z߾VśZ_/`^&.F=ޘ&8?|=l'Z mxyϫuq0-£T72XJR$.Ki^LHCb*Fv>%'D~,O2@lQa1X4A޸#`*2Jɢ,11lpLz^]Jl5@A>^%{@ЈZ,UKpanF8K/ ڀ)&("Ra9Zhj\" "wCܥ/YkCFPf!)*lve I& 3<oU|܁G΋2~) {=W>3 Z o#>trmת:wfg=nݼZn]T[]I|Rm.?v>yC6jHΟr==LF5ϛ\o-h~LOس~ox<ځh.zM‚4ß76. ˶wer>bOvnʥynE'ںmqmd%' xMd(BPc#<C=}'jKCH?}Ec>|$MW.8T$!(DB*rCm!;s# tęPA2W^D.3ࣃGRAZk3ݴT!y{EC,;y iv h=j9ϢV'?șd Tk.P፧:c\k\ "8O` yj#b=G˟Nh`KcY9t{/4䭈Idm#^Y8#zT j!JFF0USϦu.SR:Wνp VmN;(o.UWqV| w FSr C1ae\ŸCjuB:].F;h4fe^s*H L& T2뭳9!\5SQ6ќ23lL:P)kD D8.DxP2@'DLhugbg}E Ǔ]݉.[7xnC6K'G^sNϖ%d YG92lY2 mLҌQ*O΁gL+=wD(I9Zަ*{CErE&>0hښ"uI{SNB-2UĶug3?Nn|̰{g_]0麳Jf?־HztLmTno#ೄ l\pXZ5  4q0;c$:'-Ee$3#z;CI`09T}-j6$[RUd֑f;^(ve !f 1hC[%ONWwo=&P(Z9Y "r*j5L I)4%, `A>9rw8Ͳm87%)s3rz^ ^8 VG?@ &)Fء##@KvD?!lJzS>uvt(%(at\`2g3BDN H<.7F+Bjހ5G9\!k"o5t3/?W=mo A#A=(kn =~, r6֒4j(΃HcW[ ]U'MFA1]]eZWٓKdvHכs}xy&$w-hI|P-ˮuN>kP7ۭ2wlFƙKls)3BoadY7ߜ$PKH!fgY3),C.qT9* Y@g$C1*w7ȫ|7E毷2ʏkd4(_fPû hrMƌ;|3;xbYwWXS2S<8:xBˏSIoP;YnԋTj۰6wp^pNmېqFO7꒙;%)$neEW[."* gtz>7L]Z[\Z#L Cuw.;X0A[ޢa&lǼsTʴ"rV{e#^L;B%e;:׮諞M6t:ʙip8mq҂ogw<(c8W8b\[FYia&.n׃ #wgКsBNlp|?5:9zF~<7-/SYCGn8+T9i78r?#J-pSk,q*%e 4/#s@y6ĢrPB[RBr9RЖEZcƠSbȑh TےlQQdak,-BƒG,bf9ULY);:E_5(`02Ogo\bcmMB+(zd*RIA@ AC\8 ,M.$L@0e8&10L>&nYb֝ۍaɸ<u*MRtRCm$ЎN h"hQ L{^8"nTֶzl>kTc=1)!RBQ&IEyPRf$׀փG=Mhݕ_-Շ<22RV^GJj"xGvk>Llw sA*KpL C [ETI$3\h ѱΆe\Ώ݅{2^hhW# ^YՔ3&([X;r4#c0GME2[Z0rVɒD2)Ap-!r2dvmJ'HB B_G%3|iCFC_籏d NE+ i+lhHܐF$$EHZfC(\}l" $@b})B̹z>g* gwVo0*N(s3] }TD?u똹vxw3DygN]~[O͒%8qc%ԟpϾ6q|u3q6lVI73'<=$s]@qszrZ/I`v9kA:~^?N[ˆd?fA]644Lw%O×Ҩl? 鯟5)x.٧>Y2<3qTI[ θd_) 8+߿;D:O}Lߛ}%RKSL`Ʀp͵>?J}\>}oIbK_]m.ϳӧu3Xmn4iy81kHrKOk-])xC `?-J&=5L~, % }B\Vb^}Zz51cK"α sLy3JBE)Zj'$ª iwRGɭUhy)DzNqǭ$NE:\فS,St6 N|G*ЙLh(7zTʗeRL3d9+&]_#װ (YXػ=jOBuA@4 !8ke|ыeu`\ ~yeBte]2sAt"Jye:C@x^=wڐ `̢f2lYN "w]]A:|Iw6Ws`YSvu3i{ZhC6bf{~jRB>+!}Q7w>E -zMw_yߊzN;qt5x]܍457sќMtۛXc}txJ%EK[˭ adٹmsRc-]\(2k$A)#2гI>yϏ>׎ydb^]( +% +!1fdr;z%с>yC+pfTk 7N}U%LcQ! '(%7GVmlj8(蓄&c6hnPD6cQv3ݮ~v?CM`SB\F"ٹ7&#Wnә{!UD# SePi#R&g:\^ *b8qb S8L=(uCxrIgàQLACTl1ړc*u J,gkt+P']jq*U=Fv~+%HqBU']pSٮh?rAfv+-͂1ؖ셶,B䅳orf{?~ r،ecSΆ+!^}8[3!-( {L1:lC53pF" ҃ij5K`綅hH߮o0rR?iMkGW/kdIun'_=L(/N3Z+E2GC8(V1cSpګ$$Vk8䢇dQ&ښ$8VedTf.% Foy5,YHm  Ԇ$KL Z黑hH 0^72uR,JeQ,*ETʢRYzʢTʢTʢRYT*JeQ,*ETʢJeQ,*ET$-9ceTʢRYT*JeQ,*E8z\se u}ozh7Jg+?#?z ;Z[+VʼSw*Neީ;y2TʼSwJB:;>y+Neީ;y2TLeީ82TʼSw*Neީ;*edb/Ľ^MI.ҸЦˬRa̍`aAOF̓4͜~^!Jk9 |8,}2 (cNynb譯y4̣Y/v'U=* nQ=rN< ) \]Y'_Kv͊akloB@=w.ISaקYdWvTDA8,s@Cx5_i@c+0BZ`@.)t.g D4ya-[' PFAZ )Kn}LHx8 Bg.DB+,bfc'agٲ>chnވN7tNvno|jBl äB>|vP q3#ң6";&)VU{BFgdqRo)sW[:trTVAIT@.}A\A AK$ʁap6k8XLKcWMZM1I_mWT@% fR(Ќ7."-*o2V%*6tAڗo988{\}{#O,TuQZ!ա SPmub ɱV/p*h{m2JjmwX[ +ؙ ǟ/e fc $͙0Sr*V˯6ƞKw<'Ó1N16G>2K^oCf"Ep,:usC#Yd;ʀB+a+H1H8˗u6wi}U+Jn2!m[ܯoM4ߜ?,/վ=#͙^;A%U\&  XN,bbD ^@p)(|weMg/~|Zld'mu`LI$ >|E6Rb(DրP@nTHU*{4G)Nߧ-/.ʆժ٣M4i/G6*_f z!Y.Q\d,:XDKfsK *o**ݞO6=g5ߪ,eUor$tWYVS []_VdբWM?`?W|B3bV޼2}TƃZ J #]ݴVT?"azUzv'=yN>OŒG.hpxqx6~YYj#g#Z3!cЉR+J1YfHP?b6A+ZH`36 !eb&z/Fqɹxն=[m[vEnzF %$$fu)ɱ3hklPidJXs,Do,Y H!2H^tDI$a"H68NIS%7q֩'ø7}͏GZD`ZE|ce gLi$5InF )!%WG~fAꫦUM{M;h ہR)JJZ6bq[uMOnvlj"3%iTVšWVY ++ae%VH 2V *W4_|UJUi*W:|F0 ֕櫖WJUi*WtE|U7f4_n4_U1bVJU|UJUi4_|UJUi*4_EUJUi*WMg4_|Ujϥ]ij}4_|UJUi*W4_Qq4_|U[Xi*W4_"JUi*WUiU*W4_& .Z$*(GPɴMbH?jG%o~#w_2 G:J`%4/ejr|]i1Iuc^)$YOqgxCEjݛ7G']2ѫYht@YZH٥7V_ۭtA^ ˫yf 3kw\'CߴK/E:;L^G AWu5vހmWۥx@qj?]f?bzVu=_f.y% )5(oQ!Fܠ|uA }N}|;L$ض)K) vry:C{#5"!Qs6X r1i.@ e *TĒ[Y$c]gFHK "C8dR1 i59'>js,b0u1Pea_U$gF\?? _ŽZP;/+^M+Z|h_k1gY&u09d0E9B3t;MY䐭aU З/u:Єդ$e2t"A JBkIfd a6e$פɑ"$hFs/Z}l'M4J}г)+9@`v`\^om<`C__`Wgo.A%umt04NdɳMrQ1?K_Ox,̚>ǤDq` ±Lҹ~>LiE9?NC6%9X]`,yJAԁiW qgzֳQT2O.ҌuTk tUnhqrЌ'eZh6fMS n 7A$7SzXh\CςfDŽ`yW B\S=eU(oEAX^%7xnyH &2 @RF(-'$Y '@+oEiƈMU/y2̸>qc8M}+#-Y6!P*(UkR{UģEI1;yqG7xlDunmV]ۑNwZlLw%6rrz,',60)I4.q8zHZ: V5h1xE2*0M-Wx<*1_}[?Z =&4yJe%:D^C/a$s me .u),'mmF2f sk!Q?r 7qփVho` V>҃>k'×{K6_ zn/-癠w)?? 9IHoPX[OLԖU?%"d,'-+O9aTtZg*+YGgh۟Nv__AՇ= "-kn=[7ueB>(ZGXk.]BySޛ1A\0esZru]#)f<]+Gڱ]C{A 8ҕχ]O6P8t8v̌.nZcSVU?U>NݤxLMvU|s}8"Y^HIl?|cBZ Ĩm<$p(A:U?S\YQ/JEA^J`N(YI*1$e&kM/;4d[ g@rO /?> %M[>*|&`wZR-yF^mx>P'&'iTRsIX}L`{3RzKqKAK/H?8蹰2*X r*%e,>3mvd0%# $2'9etGtHrfCT 3M8sfEAUHKs+'|X#]uUxP5 Uaz\9xBej3w:!> R{+1)uͧhE+6|Ϊƺlr QWE}m2J_މcnu<ǜc8 }4ĸ|(_[3b@5Z鹎I8oW^Urvl#~|C ck} / h0VRHN'tV>{{_whEY UC=ps8絛b~-_Zh;prH뙐5Bv_w@#okmh<>VaĒil|2!&Ȥ+m'*mv'+3;(9ґ: Js5tZD~vsY bu9~m< ږ[!ȺL1afry~ZcEGσr3tDbhƤ\hq#=6:#{/',,M00xm,UBTQ7;>#B$%ZZJ[Ft:g.1C4hW5r~"Qwg7Z VI7>R}קqY!f98/ٰ-# V {y@x Pqo9^S)QV H7tq :|,y<5=Y{@I+`t\RmUc2&H2ւZPe©Y$R.8l T6g\G]iR& #3  $ $n"S2=i&Ά _ޅO- r4^_>̄î3/]y])#>R-}]u<R؄ވl6!80RϤVDܫeLJVQL28ll0ւ@GJs=kpo{q>z/yW}>Ed_Qa#4`(Vx008@Xv)Ĵ0RrVa/0&\=sU5P̕1H)T5W/\)iF\q<sEJn\)n3DsA2WE`%\q"ent!K4WxʶVp"db6ۛ"j^$;$sU`UC1W$u$s{sU\ZzAJXe9M46gz2) O'K;*69l0yhLQYs)@ hLzƛl,zI MrZ 8)1HfLo6Cwey~\ B(),>ܘ4;s5٘?;Zd}F/nC\2uya]~m>/|~zk|py#c Z6"@C߯]!KZHC[jǽIrc#%f-F8F{6_=$,alvxQ MJVƚ^;Gl"o-caCSqLv0ηHnZ޹0't 6ze!x$(Ls3|H>Houd9sLP[Z1ؔtU^wf 4/l4mI%k1xK7PLۚg"пnjYTߗ?wYZЪu~LEEL@Յa38=-V]fwZOG?svzUnhII[XǶObNί{VY[;o Ww$I_i|5,1,afȧ-Z<^"K%Yw*/U:YȈз)nz SBY@^fIurM+l>C\ ȺlpyMw9-<݆ڄYpiwi|x:| N2_N0;`!NB7է V'o,ώN&^=J;(EAGjiv?Bm er |PMriڤ\{MnT-GGi2{CC٘4ͻox1»3!-5I6l8l81t[fg&)g Ԍދ;&1a?mfY)I=E>i͏ )杲a(atױr,gʬIq8r~<M?Ȼ(B1)s*Yu !Xu^hFfl1IMKg/Ξ9'DqpbRFZ37ፕ~;3_$?M6+_ljju5s_^wRR93IKu:v9܈r&RCR"jh fH7[e%_\.^hc"1MP )C+<(O'Z҃vI7>n*ʎ QM TVĈJN\:Eda" $@9QѨtTP$HmX eF1.@)E/ [DmV|)-4]b:~z7&x}ȓ:Ԅ^rg89kopϾ|~-j=ZqSk,q*%e 4/#s@y Y qC -V)!CB'zP!ǂAV Q" TiXݚqr]X3Յ.ԝ.<{"caklֱyi'Oߴu}Tv0}'w1s&q!'zd*RIA@ A٠<l`;8jK%8 A1И2hF&J٭he[v18VkZtZ`:(Q->@mDիќ#>ز@!eHQQ  & GQEHl٭kP?OȊ5RqF䤬Fl4b7pjD)!Rb(SQ k@@)Q{H$gCaX=\ pYKՋ^^bUCgbf}jӿtVg͂u6Y^]z߅j.g0pȕz]aT֮>c>c3v AQ!}?Y5w&"+!$HPMT 9%wJ, Y;TF ̊h.9GH- :8 5<*\wSZ ,2X#ur)\ІH1AxGU:|\6+kEjn]\je`C`9Mq\⿣vz?/sQep&2`V0b~dNX>8\0mԎ*KJGi/%0D*oh/Jۏ05 A (V%T 6WK:zy)Z7˔8hEAIl` i2Q&B$C*Z8WM8scc I2& Q3⩣FksNo"<.PPՊ~]3l)B6㭽?Z]95KqdL L̹eyDCn|i:WcBii*CH ty>S}fSHh!EkKW>63Cdk;J6ޠ]̓!i(*PMJ,"n; y ]Q+HL(EfAଶڡeLGC|NS0$+FΞ|y_2Bwh`<N^ti+='/')ܰ-+3G<|SuJ)dkkHBqKT2EJ"jC%2HVTrʓA@C$/_uIPqNGEY2xy)%P-L!bN B ITDD%Ir7NmqcPȸqg1E/&ºs7;ߩ|83 :FdBwP.'dt< }w4Ѵ h(hKdb:Q&63ApI {F:N 17!Nϵ`IJ#j@hŸ1V|9IEL=BϠ7,<,8N(p*h:p$R5}Y^Lkzv0_C_Ԭ!W*9ϗog5>dW5DJ;[|<=`<~˅sba#7cF// W{I̷zpA9b_?gnAio)#ޟRut̓ wq.E`~+o+LwBX>ַ%y#K&$_3.Y=_KeL R ВL$e\Ig Tj:䭍]6ږWz}JĹ&Y߈G@ < ( B-U//UQ7Y鼘﨤c=-OXr6& -: #wiJІ)q>Sa Ie^pa0%2UӚW`u" tT=jr8O^YtLt*DROK(y*ED '4bIb0]:ftu-ԢlcKm\օ7ZN|3bI;yL˘> pK?mPV JAMXiq2ufNw! 95hf f' ZfҘ肐z,ʴJxA&^E*N z׽{ۚa"lS7S+LVfChwzrdoD=qt<>=dǤ1[o}^ ?b'o2107=p}x~+Lc`,4oόV%F`ConS:-ʻݍƋTR[/)^6cmT>ܧY o+OKtdedeYЕF(*$ց3tNZM@zx˥h)mJ8MP\G[i$46jo&IK f O! :wˉё@fa lXlq`tZe<} K˵7tfw śr[oy8v`2`^62_ *~-`.Su\r w`9.(,guPmWW\0u+BkRW+2]Bj%mWWN]}Lu?ui7F?a0!⚎ hȨjQlt0H?.e.LSn]w__^.l)LQ>玎pP6Fsok+sh,SY/'\y_D޹*rP%Z3+a`Il6&Sbԧ8a !@]>(fwvO!4]~D >K }-eڎh2 :DkRW`F]er*fjlTթJ^S4 ք\ҫ oejY[JE?h%H]!]Z ʥN]}ueo[9U]!R_O4UjU>1S);u!ՕVW8`EFd#)xbaY۔%; vlv=~Usy{(P8J=^5llBlt6\ Co)^RbenHyrnwKoQɜ79̕sd^\- s8_l”?k/qƸn@>_SZT^O'>~Ғs#ctξ9}V7K)soџ|WjR Q޳"t@JTbr=62ט%Neh(E 4jهKd$}F ޸3MN>UDZn/dd]Ύ7־<䬞DBO6lc2'q{m{+niXHkNŗc`(ĺ{e%YM77'Rt{ B)U{[6&ovS2!2݆兵hif +95g̎1;<cXn૏gf)^)CO<姊iI﫽jJw- Z6߆ eXd%m[C z>]h%M:)wIRnIu|;ˁfuiIE$Ɠt4|xMl̘K#-+iOӰxǜOǜIp?6c:s6K*Y|vhF&Eh ǼKTxJ&vNQc?l\nNq0ԅJKʺ,:ױk9lmqy I(E1o%IpfYuwk{Ņ|q'*[3vp;둹(8AJ#@P |YOxgЌp*c&k>M޿0}[W*$yK],xOސN\=B"lxYET-/ڗW|K?BjYjcU+j7/ߧBNscKb[~s"MJ>%]Jˀч8@p0:}ITІ0_ݭ L1Lƫ@ZHif_Ź嶪G\SdʖA&kɋc@|cTZ:( 0+ՖEuڌŦQ9|p)fk5:C[6k(-'DgvvʝRLI=o%T]Fr< JFs$N:! ,$01u8U%&#Xi˧L6"m`vAOhsojHLBdݒq;J9,FB[YN!c0K.E-@LNJ)2mC< &CR)4!)'uYKt)hle]m9%vpɸ]:ڮv֣`!%bsCBi$"t묋Z3L GpEuK"fEb Vq !Ld &L" 5GQ'T2<5aE0^Qt-8R"JVW"$b'7P<CgɌ>ꬍc`A`HZGe7\tZ1,Ĝ*id';Zh@"gh J͍L6Fɞp"D,-g:7^u 8[_+ye;mC\<\yX:<IDĨVGcHя{Dug 4` ;{փ?sÖ+}Z[z*5\VVy@Vi=[&;Y#(t8x{?)tVx5 hqalD2 :]Zv^'If1xAj2ԁ,5Bd(Rp! gl*<ɼTٯR\H|'dY`! ΐ5ERC0WZ̵S-g*.̱i2/"I}q`kWv&KIzO߿kLho@8U AP\kdE$d.K-Pz_\;*mxO<:QǪdi_g^Afh#._ ™1L ?uEn]ڴgY]s ]_2om/7VF7kW6nn>soݶf~\N|楑a4no~uwMA{?a;&Fծ+ԹŰEǟ>+[&){Top:<9gIU6IB IMvq D465cjlR"2гI}}~v,&3[4F;( AX bIL)78TRsFp&ήctL3YLK'NsWOwx`JJ-lG}^m|X;Wf_o8un #>\u&6/.4J${g^2/z p4s<;KWGq&K>\qKU1L\ݳk9%ʏ.(wWٵoY@l GRV&2c*`\|go%`Ҿ?~ϡGXߒZy V*#MyeiLBVhiVs?}0k*rKok]oC2sS^o(~Z>=,ui FϛR @6ѷke+eQWK(aYo%Oipړ%cVZ$2h6JB^XBsXjk#& 纻iK4k/̸RdyǭDN!,;zwHe sJ#IΗZŗzQQD9QXh1QMs^x gcC][`F+6,i*gMjƌ!IhϦ!3S7^mtrp!.}Dᩒ]xxj@ε>E0Ra(% K f5|˘K$[`K3L< {Enә:z,ZgdV{*jk`q&Aݐ[)YAK{d1YD^35}`Z#mOctxLxUj ᭿6'lh%NtF̫ ˬd7n mR.ƹ:&Z͸\o41掠Gc.1yTjfQbH `>ppM'>1.:FX}IAza~ +;irFz:JA`}9v;Lkx҅׎S_N\mT:8})eC{b$@[s6@&emT 0V%COޡ-JUuzN,dQpJi5d*I &Y%-Bl Ze+,xu y9{8Vk]ڊC@\w9 lC4Vhf_.~g*S9vLH+f6:"51ͺfmq5E_(!ϛ *Ad *& UZk@CI'޼;#OOr6E"ze.5$xǓ t\Ah1\9kdUށj3۬|t4t??Je͓m2CwPLPnOXYG7wxm+`lE$^3N76D0BEacWB ; \€fGC#"qyܙHK93/ : i>4+m#I40%}h`a6yDJ)͢VodIVR*ZeJ%#"# H-L^H` j"Q92}h]Ɠ߉OI뜈n{ 7nۡu^UNY)f:v:r<( afQHm2Q*#OދihyE Xpl`L t(4lq΀g ̤* ;+,9,[6ɱ:Os+Ovj嫫˺vb`]YofY2ɮz&iɵZ-3_]sw\B"Kz5p]rvTE7\(}EpKʯ2L_ \ejn;\e*•0JkҮ2] \!\ \ejm;\e*•RrEp VW\C$i*S%\)+&RVb2HUȶRIW\6_]]\!\*UVUR!\YB@$Ѯ2^ \ejUL\GRgJGHq~ƽ̹3aoK+ܠͳꪞ }@F2J畁+4+ƽWs,-'@tѐ['ITTDMTE\r\<}Ƚ7 Z 'm ۍ-ssL-+492-,=hHJ$$#1Ę41x\SQ/B-m8|y؏Зn)RJ xVomF*[Ө5odFUR _=JUY:ɟӐ@}e29^y\(2]tWBmKR& Z](-x Ac5jEs YUQhH$Ҙ ,1gH-r7OR@ %' pna1r"&"wQ:pu{>'G.4%8?>N\ٗ?%k'Ѿdڄwh Ձ DC>ng"X oᝧ'~Y훯q;ۯ9߁k]bjno} f}s1~Ξyt A0h_Ծu~Xk=M5>,x~t x~AXYBz݅ڿQ}Q0R1ǔ{bg=gm=͗Ϣ|zK,Ui5P21C7D/(18-(C2Z(OJ4*RUbk[9gݓUO}DsB䒢.JhqQʼn -0WH&!JJ 5E6T$@&}B@Q< f\rcU^F51#簠mTMP-IPp;~lZOb]a:yd~_1"ON-)05356r`ϯ%7(*}9uiAS (ܠ f LSΡ2@{n))(S{bCĄFB1YϢhRNCD%EVBid,Fad,bq.Xh:,x(2`5:ux{A{g*7 ;Glbm$8Ã5Fa}LʍW()Ʉ,眉($p"#{YMtP+5)ˉ ª Ma.Faa\11ۂŸ\ԶQvi{Fьq#$CLj-> c YTHl$X H*BF͐ X&#btJ$jbƨ_Qd`lr)89):Dq ĉfe@o“ 46X%)ĺh %Di1HE(ZqG@RQ4ړi*AFbQ8U{z@&_g1.9ia\.vM`񨧀%Jj,;/r] yvglhN:\.qǹx !;/afNyx9b`8 h( kMU;صʢ&ɯzetwCd~-l/U=zTB ! DQKV NbeU09p] ꋚ٨=1tra/06&O}Nrp^RaB/P7@ }VT[1 B2*Z驎7]ZKަ.y7? AlS(f7qY%GoF79Mj׫__oJtfi\r1?ɢh .d}xX9F׺e=J;umYJaژf6\-iض1T`]ƔK(n!Z|@y ӫdqWVCTbԶ>U(O$iwGnOYO4Tx 1!'Ձ.21%lRA5-)%eKuI뤭M$j20^_&r_^IO;d뙍\,7%BƲ<"m53^l:Q+"*!Q!FpiԚΎN"7Npolg<*"0i5s%t|W1r,DZ֨K|=KE^3eJ£QMypXʮG\?skY=ỜlD]`% -Q)I0UqJ$ uFSE$R,dU^TaP&n:,CW1Lk a-P q2X$jp(МD BS[Aܩ02- aP?̈́Dl :. "H@y=+Mؿ~hz(ARS\VGd!)66QrE {Ft s"'+ꓔ"a$^͸NŜd\X/cHʀG3)矄msaqFs3fh/YEkA0;G׳i-ONV~}~CO͆%8ɿQ99iԿ0cs9ϼF6Ag܄'Ox0sù.63m~h2w<_dvwɜϛ#;|,nA*{rnmNNyqu ~xSOD_"i S _:m`;$7>Z{-$ iJޯ?~VcfkPlڏVӧ}3ԙ</hd))۟+xۙכ=ʡQ`?=wc,8<3h|/.| +eM<8:ŀU﹣]b>5U?Uwߺ7bXnmFyP>BvFpLK!E\T%"ʱh*Vj*H3/dsr#ݶ=7iQ{n #A3n2\۩gښ۸_ae \[{TŵIo. kITHJ̐dL%<`_}(EEl7 Q2Xclʑ1~謏 g F9=@T<^;5sYl+;mi4Mi]scFb*_}$(Kkdt})|ϮHAHۼԻ9\Ƹ*R#6\y {`=lE?)v@2D'R9#B8K{E锨G#,Jrk)ּ{^)q x^kҝ9%Ӽ)~sdm1+|zD4nC.GMUn]__wKs!QcUI<5A*4A(Q9^u+nEGVĜ1N%ȴPD:AsygLdjyb@5RD䑖u,ghC-d85>C-L傈?D1%b XΖnEcBs59!•j<;g`ڗ]YߵYp1x3]$>xYwrTXްd 6('E` 44|וi#?;r bCl/[L`kG lg~z=TDNyl>=$qyS|( nζ9mAyVdKrKfa?~۫eiAF[xLVFR&iu)7sNle+ߊ^ z3CM&9Ѹ-<էufc֮[i?]=A*cZ焯KTέZPo:vGV[& epR¨ѲջWBD96*bj֥!?+74'+X𻞂 P7@ :-Qy "$y`b G罷.z-Yl#XN9|i|܆FwTmi'/r,c!l< ~AW3_vĈ$h`* $70%I28fyN^i9ܝ-h'|nir`f[8Abi>nyc'L6{ i\>ݸ8Vv8Q"V[2EL E78&wZ)ȿp .Fuuqqr={vY_/B^^Ly{߂F|vm~bw<_-:yqd /6x!/.i̱2 cvV7n޻ݛ=~׋R/[=EvzD|p=t_0xuFt=;35r9{nsȝPGyW[47f'yS~UïjU W5_~UïjU a;[YCV="Y!:dUU*:XV*۪Y!:dUU*;ّZKZKj-ZK#L9 F7 EџsmsHZѨѲQVpelet_>:}Lrr}JcP~ 4ڔdTp e| 4*JKZ:g(OUWC궋1mݶAgq>9j:k^}N9*2ƇsULI t>K!h; R:J;1 ^$ bE B^+0T`x`[{l@Z=)$ad:,ԑe0i)"q|w4f?Na6"r1CT{*Yr ClqZ꣯wΖ?-) 7]~܊bcm˛sy 01+ֆD2q(nRFQTDFTxh<}+< ys1qbc^b,.~ DqHմ"Ot݅,1V89Jf(#`[åFa-!<9DFsobZ^~ޥ+tv[Q8d2SEJ[vC0gt_*YF>o4Qg="5w.I#)t [K,h.c¨BqQa;#[vʒ:;QU~+Gnỳ 6y&D)8' VQY,OmBGE+B-"sB{(9|pjGOJ\3/jsHbtQIj,"u )B\ zE@/K`H LD1L"D6eAj?񢾸j#VRY=V=ᤒ i|Ozx.I:Pٷ6MrcgۤlV^ҷ?^2%Fj&,~,+iI X#8R4$X&lVc,MFJ(!Dt|u&lw8a+^{}{;0wk̘Ϩo}46t];9;Ūe?nzBzS+~ur㼀n5iR=8_%3T2Pa'%[\ݟf[ͺD["5WQ(~]jn[Q@j#ZjmOz:x35C %>b˖/~0` !gWPNRؗ'}7ק,0ӟ`jfO,в);')vlFȶc"C\EyD!J3A K-"o"r[׋m^j(|x}Ϻ%=! V i \;GicB16*i*W{Y|Ic8;1Q;ܩ/哨@tѐ[gN^IEKD\%JeÄ4P21C7D/(18-(K$F)RYL>L+֝-S&c/IҥJez{J4'DN.)D АGg(DLBaj&|D[diSl .ISED$qɍTyC fEw6_t\W㱜c95%MW*(]mo6˕>=D{էJrEWYGQɃN)r($(`^0. ,QF$gGWĄFB1YϢhRNCDр"ӪPug32*Űd슅0ҙWL33^Bl=g1;8 }15hdX nxg>&c@V6E=P:3QփC6\d{8˰ɕJx%f6" J Fbٌ~tLcAbұ+j¨m+jWk;`D7Brج"vDDN"bWH@Nh]PIiKP&+<0P*HNkaUh@lPBhZQ 2hP*9ВF{;ME #blFNG\bӏi슋0.ҊI ,.4MDuXM w-# ^,A:%3Si-cbұ+xv@؆+]vܸ$p@eFRnW#+)j\(G U]u|5;Yӭ\CBc9n %E}&KSFEL1Vz#0獯|OΞA > bMR=4{RݙWT Zٍ!4o)xt O'W{fSlBX=c0P{1?TlخޟO*}x;[Ϳ`Qw{F\'^~Op^1@{=([K9O;/{0{GT4$P~w=jJѸ;fR⪚ZT#PK^)B*_xlAARApֻ_] vA/^A/9`whutfQoy{ d[5lB+Cو]^EFR,x#1jߨUN\0w5u)QHBÃwϹL*&&ꣶ4GKbc<8?|/2ΗD"Vx6VuWu?.G0uIM<⻨rZ8K2LL#\,)¯qJ$AꌦHI<[CI_4" 1~ڊ)xLg*Љi-6L3Ppz`0#8PbgnX(i8xنaSO]?'Q^⭰(0hb^eY[F,?> ZKN)JMSa ZJjM93XJzaMy%W#0&R%W'[;zLF3nSo:K2I&_!`Rx_ SomarRs5?n8Rqc( }+#,Y6a gV ȔF"IΗڈ/66&^tE/{zTb#JmӉN9qCnq) I\`(o8<t"XU0cx!w$ÿ^npwc9Qnpw7j[ M i&9ʐW^S~bhX @(n%qiG).E͂,  5Pt&<7}18* SFmƆwHWx7 4gzG `){/9D.e >NFo_{4!sP'.ƨp,B%+5@Ƭ:u=7ؑSai^g[e#) 6"B2N*F` 1{lN`0%# $22:pp L'd$Es9]gx Z\ i=+g—1u~הa;ڄTSZ"ӫѫ͑ţD=(mGhk@ȇr Q]adF(7ҿfĘɀ4~sɅUӏ.dNl9Z$*r!IL˼N^K׺^>_rɻd<Œ:(d9Y,JhI9pV\`c`B/>jMx,Tk,kd[,@kpQl J:Z4 a-+Mk $A;]JD]GZm]J[ׂ Q_0mlҭ`kmk0Cj ]!Zq[R!]YNT-+,_@Stpk ]!(uЕܱ%᜚^<1]ԇCNd Ҝ{Е]Ob05tpm ]!ZΝ wHW -+lTk ruh٫+D)lGWVB>ȰcoNWR;+!$Etm gEGW`xR[DW؊evC;W{+i]!`i[CWW]툒vHWZJH kB5;չwHWF#DWش }۵!1(v[;z?teeM;XU-th=wBsK>]ð@> zNj^Ӆ!n&#g0\~-cC SZ)!B1\|>Z^<}a=U 㐯뉎,q i:*jԎ PS\Cf)9?>KxsM\) \mUa2E3dJt[]3MŜ1"BU,AmY:P.Ӑ6V=n0295>&9&gLeO OD1ebN}S` hoH`nx첱-)'4nTefZ3U`wrIxò%$0F<{/cF'R+ ӛĂ˹,,ZG]™depfrr& D&U5<}Y41f5Yk2mJަ22]6MHԞH,94nY`rǝQƨl2+}/7gG(ڟJ8WHO}^I^_̧eAZ׶yɽhR26n1[o5rkB[z]l;u"DOi-/Ҁ)iC!47I ˘B" +@IadNmZ=oImp֤eNj Q` ,M(b93"h%TInz˱~MzỾU^,O?_-;ez"ƥ^͍~&,h!#`R.hY(+82lI{}v'Q+7>/qmzDK:\2ωS>PY% am:g(Ou7 q\iƼlf>?}^3 L`6C@G'mX >Iv_FY'_b}T\S!m+o IQSMƀ-: Lދ܅ ᘸf̱uK&n ݬy\ᘽi2d J_~~9۹Q<۔U|'ШsB KQ(he򔸋&}T^DDhp@tzb=>{MN{Bν/@fe؟cܧԌ%sD5M$1RIj,JB @q :WDe jL(V0=AvڸFKR( D[8Vnae g(<=]>3sY)e#U7Ѱm;S"1H_oX61񷆋hdž0HA߼^Z̘Y8+*@U"jV,qH6R[M uA d?Ay-2jyP.Pَr[c/{K-}Yݛ%Nrރ&^Gp/f]dfc>~\l3oݠ19uU V F2J1V*i*W{e>gΎ-yv{=v|&0D' X5 uB{%U.QhUr s\P+u@ 1P ༷T,NH]rgW3U#T -֊cr=Q2i_S9!ZHtrIQ%U@8hHTqd&+QD$X%Eі c J-b9+z$IZȨRVtQD@3A2 ?pȝA+&8aco`<ЬRD䤬DIN"x 8̻Җ M2VxaT$8%زJ(-6)8Q #РTTs0!%$w;z,?DGXYJv\\:w4ţ^e΢ed\4&IdpFs*}cŨcWy C<܃ې5fB7gFoI\pV@Fdlץng"X1oioĻO(3[MSbEÕzF;{yDgoURC)ls/O|y/󽾦 uN1I?y>:s} 3-qה2;htPQ򨭻]r,'N/>zQg ɻTf3D-U,ey,#Pwf"Re+j3oF; m }aqLs3?bK@wuvgu.wuecimv|s=z1;tq!|Ssl_575{CgC]Ƕ8>v ݖBȆ]yMQE|䗦sD͟:t<[{/ٝ yi/==Rcf4/@U7P.?zK}O~ʤ&RʢViH]g̮kjT>Sn LKyU,!%Iv~E\.ɂRi8s:ymTAu_E]wa+7sfvvnzv o?u@\OT.mγ~0C&jpttЕדPsHߚN(kXn+\jRVm|S%nM +kqC=i>#ϕ%^(?CE!)]d"h=3Kɉc:Js`bLLhRMP.r#+̬P\&gN#GLFFֈyTV6e~ Z\f毷.Pmzӷsܲ8*Bz1< 57m5L"Peb/Q=!v=k;x @&,Ŭ6ި(uK:&o{w b'"jEau¯PdʖN!<.#´dRP[ؚsCp9姒疡w2wyn/0-[sNb^я>/| 0F޳Y!.bMK*w{k;ާ檵e/95Ga\R}VV~:ruO|r]bMQHZe|ȕ1OLĢŹzڔ;'<5eԆpe^o3jmmH(2iٛ67/ߝ 7B[OғIDr{tR~0JZuB VKz -?Ja:q"ŕ1ٞ ľI?49gmYj)@BQM׃I@B֣oo?ZZg!mu?˧-LgGNp sK,,[m sJ|/If>1E643dOm9I$[uLRrIbaucbp1 W픫ۯy+sJ]Qj2v*Z DMf:&~m!mIS:EM7ݪJ@Iv;3֩31.D,]m fe ~X ]8FS0^mM]4 Zs*\ZB+j9ujo@gW?l-x|s/|w1r?~ozo?+یثY | Kuz&) *$3%e-|$oRP.,Kp&h8g,xT+fH BÖ-{x0Q|0.ңf\оky+n`M2Ũ,LBD !,:aYqC@'~u7w)'?\c@Q}=ix-0z#^]\,wWb7mto~?6tk_IM]>ūM30yPۻyU?d7 B.tv4yY8F Cb$&TKr?O L)*h ;CQwI9)(XLRLU8|/!#uJVY$G"/KB!JElzeT M jZBF2*%$UK@+ jlT6l0kU1.Ul9K0&ގ _/RUe+—[ <+LԻ?m+ hw'D-*Ba9Ŕ)edmD!D klI`P&;NIwEgOg^!;@ `\,%ƾ m'nC\G^o𙔵(:aA۹`F[LZN*S+t{\;hOm1 w9qA%-K%KtW0Y1b1ĄR$uZףsFl۞P'Xtj+RQ&Q,|d54-g.y))CLe<׏뫴6s^1%[==~">_vv6J\<b dm"QRlp@ B0 Cv)\Nѓu) 96l.osv!zdP6{ ##m0 w.7[q–|~GW:z{n^Qm9q -x [k]MA9cGx+0li<ۻ8@FkC#-U&LdTP_p z(/kHH4Aj|]JA6D5 y!W#)xm)!̄KLs,y yiA7(iAg?Zڵ/2Z3{yA7X L:?S׽o~&Q P%jQΗh;CIw;W@Zi(QjujyKz$V\He2QO51['.R&JD̓baB RCa٢@H0ǔ4be,)M6͖gV>%w(m8Ғ(dD+DKT"IWw˫p_XXl7UT2%XJdXYU({#]5BXQ-gқJ[ft[m}sTjTjW>׆]%2yדٝ)w3\7\<\E{Uـ1DDVg(%r#4d-kdcF$ܻ <Ē՜X1P[HK1LfUn͖*Ͱd1 wuZne/0N>Od˟+Gl or1 ;$O.z\(SHAkLKau1*ًNIiUaS;$gT@6jsP d+#vF4<n&vha@")XM-5hD&CvNN6D6X&9y+ 3d a5c Lώ G(Ad&16fٍ=U`<@Vq "jg 8 7E P)fW:`|:ȹl=ErpV@"-E"m scF Rq&a0t4I3ɵzcDl=%ĸ81.\&Q |d.YJEB-'Qdjalt(%d=hiC21pTt1#sڟ;gq ~10>6{ !Xef2?h@l,wWB?|!汙pLAbθ4zFWfa6Aons]Ӽ6J~.a]o+&OvJۤ z1m(O!1G!G~TSW"g5?Z@R˙ɶ1vKܺi&VChww~Rm5n:ao|j|n{f~ܢK|慑v<2 v^9K9|3C#)Yxk9/_X_v jso?uW37 GVሑnwk~I] hq a^8!tFfhrf̗GA$!Nx$@ZFMc29TTS9Uʾ^f 2IjN>d>im!) P [@k=彜=D1u%WL?M9 rZ[GUܰ-K3|OW~X)>QX݊tJh[jz:Z :(AɤPŮSl/:ҭ<sP )˰`+0 C9$FPdE m0nI,K(74UcSRVʺk.?A=\R0!&[˘ʊ I%WX R; {Ey鹎 -H4> PfyjXKLvR'Ps"ReY"u!G'$FBxP:6"Bd)Uf{֙?h3_L'/{ Zx}ۤF]/&t]՛Quv=G~O9;ty}S_u7 mo"}{߽zUklF8g|W?N*_ ̿hlq(mx}{Im2A>@puSsuVS[bKyS ?޵6rB% "`qu#~KKD"ڍbz /"m?9aMwU_~hR@ X()> 'P@r%@TgA~<,vv}}ׁ==wbjx2߲lC;jmuHCxVw89MNm:n&}0V,b>MmZ!7]z#ENޝ9· ,`+Z@DSn-+q~{%MF86bjbb`B^٣];Xkϝ;Qŷv9U+ѵ{r 9n %u4qZ0WFEL1Vz#0Uw)׏0/X' fCVZr{_.`؛}yJ.MEwדUS_/Yִ8|\VS~`\6qVXR x;= .5 Bi'HxFVHy>M%-$2ƃq :aR ߼z[x[c@}KH-AP΅#|.iMp2W)ۜ`n } E˸|B۸af9–WoXw*_. u*ܸ8V腼#A-YdG?NZIq.\U b)9"rbihs[>jzH^f6u^EPm*6I`꺾DCz92҃BKk pZ܏G!"tJTģP%u)`C{5(sx1wi.^!fܹ~,ٰ-[~![ Eg'/]@K &)}>sW/.)%R14&ʥ]9gg;~+7nbZ'9Q} vn~mTg.?K69%Z6 ΣL `Qmt-T"_tdw?[5mष7pE;8)R1pYn$G8lz 4wlMɁNFBY9qʧ@*!Z!sH[]=YptL6f]} w~j:o}|K=*|ӻ[gch11>IŔ@璂bH& $\k QKB"VD8"d]q / AykC";Ry/M9Dx I(K-(u$BiL,[AH; Dsٹ0 6"r13CT{*Yr C1q(Wx;Z. Sy|lU"3Xl%«ּ~ox܏0ycLG41z ht9!T%g2fyJEo>*/"~"4q XO׾&oo_PrW<ov1z5Bx-]&]@AR:wK̺!.A"50f=o>n9UK;SPAU{pv9ע£ǪẂ{/PRO_)7Vwy=ﺿEnІyI3_V`\;Gic16*i*Ww`gG Q'@~-T :hHǭ+r2@or =<.˖hdbo^Pcp[*Pg:Q"u;˝^it ;XZ>C彽JuD :DK(U(09>C}%d*PSCRm1S*ddpO"8("G#ьKn\LL(!YJ͆ :R7έzn%J-'ʌ9oFUP2[;v'+q}RAO檽"k9U!ŃМ:4 y)4nP3 &ty¡3@{n))(S{ 11aЇ!71YϢhRNCDc%EvBid,&fd,b+Xh*fQ1Ξ4gwM]~ffƏ~wCїA[?3bk#6Uq6 cOzb(7^9ãʦ'JGs&pÓapg6A DFt,'&&,RD68ۏq)<.-ڶv`,zhFH=65[" "Z,s$DWdM,o$!gHFсFfB1 ?ph+8 Uf織~숈E*"VD\ :qwA%-Aƛd$*@ 8 VI=pJ<e׬(-06)8(Q 2P*90x;ME #b1q6# y1:iɮH "Xqq;ROso,/1TJTgV)jY"3glhN*.iǮx ! aEu5rc6UlǮ-vܢV?P$BCU?fNd\P_E]Q󳪱^F)o 40b=|R$r-@;/ܛ-{픛 ^ Kq{.:N4 <2bV:.y\ z8ru>r :| >YB%k|Cӽg/{WƒŸ!H$a~y]- H#C"%1`[4{gU]GbQBfVk4n&dgD`EdẐ3%e4]>+=%`rCQ@{!א}(^)ArUN:os0 \FC]Wne?%$q+ Ց sʆAO'hJa!;t6z1*JFf~`re=׻Dgu+k?[(#(:,&%lqS24G('JWԽGG5ʹ?9fb5 |Wq}ׇvrwshEL:BDH@pIi^zPʠ(MpYs ۭ $ǷRrm1MkH旷)rrwK tm,w1N%[",SAOʣ"w*YϷ&HB7[Q4h'ѢF9IgQ:gH-hR:ZĤ@$ G2Uۭ Ga(b THb{Vula݅6W7:$NƐN%A(byA(kGz^Cۚ5$H8XG1"%y,,Gi.`QVp{d:N!+Ruڐ"I!4a栽*hsc#09`KHcv,ε_o}aq)) eGun{Ww`\}KoƗsUciwwӎLuf6]~_zSb No5T^~#zo'7Sk-]Msk wx:Y_t gJibaC2䇛aK&5^rŰfƿ̨Uwo>_Թ%|OsuVNz/5 eGHOeu꠪~71?6'SZ?vgQcI=}&N\/h}"I}cc& 5fynbC{7>{oI\j/¼#R2(գz+c  _!P\N毘\Vbgr4{3 C޽>KB6b6NPȵ_SMu&Qozčod-HG/o^gWinT>qڢt`>딝:ˤȁAx&:t99Ϩ;p^(~kW~*>'.je/٤"'5_+g IgXeGeBBi 3/ƀ+.-XM< {O8+ѮJ1뭑ݖ[B&QkqlLy3 6%KSo&YjGԔ$/nm¢Yr\0m\a4!5 Y'Xx< Ri]b 2yAg\32k"aC@:$Ӯ@!dSvhM{ A 9E F! 9"T!B@ږ`kxJ5+4O T3gEmvԳ]xkuGFtmzgDMbF3ݬfћ9ot1fK%l~̈́4yݧx|7x3?3FW7YܭynwWL<ſRj?ʪ7,Xs]CWdgO[sHc&.=(]˭͏dkU4g}!5_OK1 :yn3EZ,T֒ ⰳ;O>djb!,urAz(ˆ@c NhZ9! ͖4k gV@8C˯<$֨}>:y$i CFKMfL_\&mf%W;Yeb0NRƲ;W Z~83\ Z~B%hÏ*Ԋ! e2R$eĭ-oM \3Ch MMFYнlDgSQ3먉Ľ(5c3[#g&S  g%\#FSG:?<.P\NrUۀa3T]0wub:z$j w.p¾ (Nj gJ)`:B̅Jű$v}=weŷ^n u:nȻ+1}G/AͶCZGBZ^iMĭ~4GH̘IimhPB{57 GF ho;յ=n[^ h HgnP)兗,5*,3 bIKu[_:8;pBmk.gh=x\r]+usёQ0鹀8GlL?bӺ  @ "_ً6   Vڈ:|RŻYKDEڙ1)r4|1TH2]0uwՎ'!ǽ'z(~O:@Ίdtj2utT_vwP8POq#nBBLIRgdwgyfoJb%a<,EIu]ƉfhhsN^ 1K-7Q +cCvJ\B")/] EJY$욷^k팈n 77>z)`:f{h'js "E,@BRUIJ2ȴN Ve*)SDED`R1@'K$" UIm-]Y$O>Nʺk/VMӄvEgY˲ ^T%􋩒^ɽWīS^VIWYc{?eVk4nmlD`uM$͌f{ as=la!]}wxY _/,~VYG%T #[z)oݎlLcT .Ju1Ǎ :d|ɱ@"`TLQD.ʉ"[u GڔvXxfj/ )i6Jƾp6 Ɯ,Q.p9۝0}uknY]T> R^xׇwsEL:BDH@pIi^"CM锠TG6#iO17WM<{}Qѡ/g/[ثMJʩ8cD)J䠹9p ih36CR Gct -xJ*ڛlf2֙si8k+Owe8ӭIFlpovø_>U9d|J 09oU hO;g8|̝Yz:412aZaOf1V܏~J ")ȅiETSJb(G/QQhk*jѥ 0+Pr%'KHY 3Ǡ7HTM25swBFX*F-KNIDPUgNAظ*r5hC*b#ǿ2/f~_r#.Kmַ_rS"zf(Q_ilY& 41TDXVQqCiZsIN%ƗHBv Y$r7gUe3CH-5g; :9D~&U*&f/Gl=e:_ [f 렯(] 2[up쉩\0kJfvڱyhKab7"g?s]^T'oP n*xxEOK;DIwbpw#$ܮx~6\WYBFqX5t_:dsd˜Xvvd76"u=5%??&*M UGTz:/g=Li޴sLQp`l1~uFVް:N;p3ƪrfQ~2'oJolcn@2lZV 9^4sV L; Df׭SR|R[BDx@zmR:FA OL\ Q:D+qpӴt½ T&t]sZN'OLWRZdE RJK]WjsvlčW(:K1XOrJI!2A&8 5<Ԉڍ U1HL451sDP F%S69k5IjV okMQd`ˡ 2,SG(Q&ImT*+bD{JN\:gHDH:! J5Q Q"i<?-%C&Q 5PDΎ!ڷG-t3j. qS$0_w亪Z4a۷irVdp-ȭlʂI{?SmU*z{PZcs4(/W)yd(cxʻ|maQaQ9n(iWZs$7ZFBd 6XX (YmK.UZQƾP, u' Is1ʚ7gM];ův0}/\bcmMBOPhT\;i5BIڞ p{5Oː=1βRy NBPT '$&D-K.hLcڭQǾR۴,M'; v͢c\Pc3*!aO@3F/%Q@$3l IBF͐ZўG&DTXdKTQFS "hTGen:aeϣ,`[d$#Ä=5><"YX~UN~]߂`ew2(ܙb/ /5۠`s*N\ 2bFXp\ h\zV6x$9qr>r :| ݒ٤v&30ZC?h?eE;v'% C fLP‚7dZ,A3:V:<t ~_$ά%tͽ,pH 'ƥSK<şىXty$'ݾ>\8scc I2& O5ZkD9Byj;'5tk)R'X;oJ}M5=k=?+h!ة>ϤEsAjW T* JR MKB^v;TiuI5*6BHn*5K>)]RQϙEV`wIm'SI}xT0F[@7ۋEr#ju]9Y^!קm.(upUYnm<9к}< Qr An4E*[€D2eJ>:r[Z04Г}}~4āDg[4J`\pLS'' 2фh"yJ M+74(;KQ uZ&<{Z )S/@DC`+L*&RݪH磡M%U3Qˬ 7dhrVqVhGL({Wo_bnnx^ԌȺ__.~P*|@JI}'[eZkG>H4m]a_Dz{+ѫs/S$3% SAøxCNۉS=ǔ\j󾿯LdT;뮍MA 7I^=M/|G\?FZ3qZǟr}ԫI|p}a[7= a @ 3I8bV=u#7l_΍9;׍T*A6UhwV>[4CZqEY;N-&Ybv^EP*P[V)7j3HL y/(v_%]/\Hjrx2Jd [ziZw.u\^vK&(&/ӯ&6:N5P*DxL&PʺJ<2/YȏYh`HAJ2u,)I^iy) S-[u,i#y٢їlz6ia;}z:DI!%g/1P(;0@vUhrKI^nkpڲf~A}[e}Clt t>@3 xײ[DWl^Ϗv9]^-w +<|o@F)dkkHBvKT2Eʵм_唠qՊJuȓd[wFTI }Ku5}S)%P,LW EXȵ7U"#DEru|=+B½&|`n Ys3.I$ l/Td I1CD;!ljj .AGB:vb 1A[$1%%yar u&.)Au8Φcvi*hN^iE}҈0Z1n!766'ORG[] cv~z} Sg3\h/C)Ի4i=Ưirڿfg׭(/5D݂\g:܄+~?GYWM*9ύߪ~]a(>*bv9}cf6 R T53gVW ߏG]QyI ? 7vUjSٽ}P9sl)N"x,~c^>&S\ѫ [ޏL8-# ; 99N Ϧ3Y ,Ւ4p u+,oEH)T9-"IfWbB,ZN7yj%~v~f\Xf͞TܧD(KikMxdށS, d1RR[Nj|,]Z,|bx{Uףz79kC &C _#wa6 (hP'N&C'KT2t2{N #6yB*#Ɉ1"[E屋%ӝz*{+>!q~ĕ⹏񩈫-cWKA;qŕ\HqB NF\et=qG]e,,N\qe4;!qsNF\et8q.W_;n>I@:~,g5wYNP2|X!t Gaz]yfu ]^wsZRPZD) pDrIA&g#&Wz& "Ch4.йO92m$׿ x睷|kg(TkzB2"~_9Xf!h%W?*y &#PQb@%,8c7խ/f:5doz{,sBX?{FŸ`G" rE.cяj0EjI:mpU )DciLW}m38n_k_>K/k_6G˦ߗOzjиs*x;O ;_uXtKiQ i..] q;b񏈿wZpT#O oeCZuVmlr N&"#<2bA5y*K9L,!ۑ䍋K)vxߺ= ((^sd&bR&bR}+^3MYv%JD5ގRF-iJ X%8RT$X&\_zaS]3*駼xrqc( Dh,@{Ђo[1 B2*Z驎7hU+h--r-1him^r= F qp ?gsf w7ss_ƟǓqGR}[*rׯXUSPLTe]ӤvnJJ*!RTk~՗$+C8^y\(a+cJ=GZiq>7{I-M)hm6)#QZR{aA4Q4kX_'H|+i[">b ΏE ?yi'}]\ Hou$ qŐ[@mKZP>M}bu5^;gM]t跽}JXõ83vyb^؅5<<ˡ40MwP޴6%:e)JRh u:)Qryp}6Jޔ.IU6ʋ1 vFq!M7s =@s?R!i0t$E`@ q!jI(A΂Y7#z`!(omH{g@* *~ 02eQZ QH2h{GZGG#vtxj.<2CM>\e mVcMupWo 1+ֆD2q(Ñq(#$ѩ {)<: ==|mC 8 z'BduY_QLh PM{)VvE_/uբcP2Җg p;) P&3̿rn}ƏS:xPNr41 IH jՉevg+0*(;;.pƤ gk7&$Zq yJA#Sp3FVꌘ?oBAQRvş-'7.M .M)9tآ|d4 sar?pþodQӵaP!:>0$Ig87閷ER<|2}{I[Nno:EHSy]$#V?$-o9|aLZ u4ExgGӧ=e2ҳZ \_tv4D9 u)~rIuڄnz̪Z4rINV< Xd\1Jh:2gW*!ZOυ$! $h^IEKD\8%V/s;P21C7D/(18-(C3Z(OJ4Yb/OSPϥ6,SOh"%E]Ti'&g(DLBa St ڢEp,PDD3.1*/sr 1B- ~5EfD9uiAS AA0/c  N1&7}=L0Rd=II8 U,Y ;R IƩXh c4]֌;ܜ4ywMv0 Ѭ7*7MG+Glbm$Pʃ5Fa}L BQL+gxZdBsDSnx"#{YMtP+5)ˉ ª Ma.~q\11/Iǩm Q`w,zhFH565 [" "Z,s$DW6(I47 Q3A#D@3A8\As;6Q/乔|EĖ^="vH8̻Җƛd$*@ 8 VI=pJ<eZC QZm Rpj!ʣ,e#РTTs0!%$w|;Կj8[ď_0.{\%&kK4չ"5Y,xYglhNz\<.^ SCv$E5^&pJW~Ҋ=&K#/#_oztWDTB*_yDيFH>v!'wƐ@<;Ǧtd쳄%{7s >) MXO!dDDKZUd@୏$jkR1F틪DaRk(c$r *g<T&Gmi1v\Df. apk78^\ޯv"+]g:'_vLz;|$:ڥX Yp%2% hPF/qk1EHy Do"AFw_8hM\5tAhô:!'L,Q+.P\v`X  'g/'.Vڋ"$`SȿysV\AJT\WrErm#|ơߚ^jk:h!E)@NѠ6YH@Jh9*6J(Auxζ_He)V '!weC}R@$ ֩e Iv+@+ L}ץ+R4 ĺ*̲;k8e |7(Uqvv=fdmS!;}Oj~g"1~Sa 6՚sc~7p8 q0˼QJΖ7dBO\q.8F ޿"L,AQpKg GMG5>|0GSH0?]eI_OSAvgNIqE ZVG%vuK>)Wn6?zqGkFp" S H6]ɈXYv~Kz8ϻjQ;gu5*#.fYje)Sf4\7¹g:YׇAۆ=hX4h7XL Y~5sCIf2qؒ tnOԪQN=daγ`V8nQb4bN*f #γ$׫@}-\[ +9),=ٳ|3L i'$Ekʕx|vyMǥ-PVCnD:ik*@=d{Fy*];mMe'v>o_|ۖ{B;m ҶoJR3o_F%$$ ԩZjηYǖKc{;Sۓx)JE/DT[! .ϰȝ)oTY >Jiɻh *ulKױ~SClMiffnwIK|dr'kM#]u"g3B6oHYsp.[m>iݮzP~\0Lv[\dk˳%ptpӄ}tw,Estw9cyc)Udז[om哯3OHod1"b*kۍ+ǮRKf0sl^6!3f: MڊeI-ɷ,*I%YG*nvC[*Y$e-^6d@oN>Lt-Z0%d:;gU2Ag`D5Stݭ̲B=\߂Jz2hz nG/^¬Gny QGοsFh?!P6~2Gmr4SKw *j-Fcw~}'w"it|*T@2ބ\R6-x-ɵ\7[#Ig% "B&T՜f!q- 5zĶj┚3l@tS?}~J;^? 9x,gG'w>\IVspo[%JN3:RTUqbFW஡dIݯgjrơռeuf}զDQ1>5dʈ.QbMKfvkuJrgޱ)O?tBՏ}'ukuffe]Geٵ{a˕zt=IZM|.hpfv{K'FDUOH'!QS-eccW\Qu{W+1UNꖙ49ػjҲr 0p&:ƳkΑq %NHAHQUms_0*.Wk֞+޾y/tMBS{δ<$m>ʎαo& ŊG⫺WSƲ^ʸ1r|mY &$9YlbS6b6Ԥc[6a 7饯+ZC:0^z)lt,%ˬuQU餛.*TrBU:H5)sd_[R%mu5(a /cus,>S|Əf#xX\'(㧆~96=8x0Qcfu>Klvm?፳OީռkTcGĎ,{ Ǝ&N9}cJׁ])Zs<(urɤ #X&:2ˑHbΠƺ5H/`/KNB:7yܯC̤bCmA,~ſ_%ͳu*# s&_:uM '勑ή9a?%߃:]F[ˀaq yŵb X_EݦIvvAܜ8c mKߥ2o?{"(/9~}W]}}%k,c -׿xQ/Yi]a;Y?U4lœeXs5olW|m.W[yڍvg!js~h?{.\ s2WGV4Gs}q$~:D(jH=qեq,~\{8W[݂-o0԰ o՘&՘Qs$*4-vUVi4p\PA'oj̵WN@F9 h VlnwA.9cTdPMr5JnA&v oχs#1p:{yFv";wPf;*7-_,ahoDx5q\;]4Ҏ;=%hrʤZ%1.ԩX(qƅ"LU4ħr<ٺKdtǻ`ZY+DuVYZd FDU4Cbh6ѩm.Dk-}! @l5dHМZ\dlr5PtUSJ0- sOR7 wpf c3kX38fsGŴ&4*rl%|k[@Ţ5")*M`6fTS$aõ`C_R5u(ιaa<&!x\*^l8[H<]Q~nJ=8l^W#.Hw zzJdŁSѦ8u@V%cNs6'! z7jn䍤jRR#ΩTRj6j#ȵs2Jn :|yx8ǩFߕcnN炨(/  el^qO2d $XΚ-2) uAQMB"n hM"b搃)duAڳZfB !["Y7SIeƻ`4ZE;Hs[rĀE F+A^kOuq3 R\ŁV^zWa Xty4qӆo5e/y ,qFθеoc%i̒ SqV. U, J߳3 V4zdC껈Cm ,4doIU(*hMA%SX|/b0'؆~BVv9VjBJiRTLdD6/ fcC8'9 H_ V[d;/5 hWNw|B͠j`o]\KG2P&c@C='X "a(2]c#5OQEgyLn#~$Xkah (bu-S229`*B)Sj]&)@BY3'Z,k u{PH&! k }Ѥi">x>9[= Tb1k,8_W !PANNZC!&D_t &`xf;/_~VաWgZMVF5Vg n8"j0@ƫG G\*`tY6Rt%yi,Z*#h`6y@G a _4+XH *x$ D&r2)2ȇ`Za0=i0FƁ>S!^8!/:Y`G- ۀ8n UfeSYߨ }^[9UmR(؎ҧb|VW}z_#O Jwl2O&6Hc c.uH96h@/f9ІHT*衻%B0#L˶]J#e`7yMsuĐ-hg[R; yQ} A 1CCx("4\o axVƜuSIA>fvI516ʉRq`k!7)/t3,vNTj$k4BYIx mJIv}MTE*",,e$Yn%~D\P_ַlu$*6vi`x7:r7gz۸_!\ͳ/0e`#KFpVDjHJ3uE,RH:[U}u4TǴޠ:WddɆ48Rgpd;EAq!Y F  >Phۑ%YuL0.rꌥ,Q0Yi`l-r~`ͳ!-oڪ'y\pIv,x=L=R(X@PxYAv+k E A+Dp>_9 g #:kcprb%o#5:*0aF8_N(+(ư $> 8D 6lzmK`X\Yђd>'21VعfO@JXZ= l$f|0TKuUĕTUVcWWEJ#P]uezpaL;+,{ʲg}Y p{3_-> ^?|I΂V9R &E傖(vg[wx#:+1Y)]$XP:yQ7\|?bkX@=ۧgOx:)Ǎ\i%#*2W^sb4Zt"iAC,g}JI@}ꍿ+>bJٺ\Ȣ~ n0ѼEB{4͹8oܾ9u0=oξwc=Q.!Q(G.ͥ^[>N¸sY9%u -6(#s`G,|AI_6ff1ְ(~j)'t^c(xCQ/<{: ;T5pco|nFe?݅jsMNy,IX4ۤwfsh ZL0A0+i\}\WQm/.x%ͤJ8*b zQ)/o~.QSР99P]ev>r#oy6ʾ{Rcydk/w;޺zTHaZ ;gQ:~?QV|LR7R8$1oDKG+uGK-^p).ܷ 3ouyjz05@Sc2;cxte#@X|t:A ~l׽S?/hkI&aٽ_=Y5^M5@_~ã)zi&ܘ=9npu{vy9^ o1(@%ZVja3Br!z9<3X^\]ʿWognݥoO ԩjnZޮhiަPY;vZuäSwxCGqnxOKf}g>L\j70i|0ä^֞ K*:>2*Hf2%|{/䱫}hmȧot=ms@K̩TY))༳Qsbvyf@5er$l>@AĹ?HXޅR #ݽBlvAYqBr}ߝ|QN7`wrذIxò%&ٰF\RS:h nݦ$]ш+%!VwUT ,Wgތ(oso(-~Dxڝ3rؚj5Ɛ,WVxxNQ6^֎(RR8Zu Rdk h<ǹ^ydHH6004nW7nX˭y ,'" r%f2Mk•"|`J3">P2 5D5i ݿVܿQٛZ7hcǶ/@Vzȶځޜvs}S4djﯿ=@75Pr3{\{S+'ʇE5\AQK2K1d1e$NRǛA|Mm#XyM][gXbwǫ+y`Jѥz&JYJ͓T+z߂+<'T"9^y\(aJ!y{/9iZ>hRQk>M^joB&,$qԕdia"Ub:ei,ea.ќI-$/u5OH1 k9LԚ8@'K$%dﶻ p*}3lVro>|sFßk'LXamɗO3ɉD#,bng"ؔ4nV6{{YXdڛ }/?r9 */Ubŋ]A:[[7τoPu?Xr2Ju"-^ݯHYEzQc' dž_WJaGGa~*쩨ӧ4G% . e)Jrh$u:)Qr[G\cB0d?W=0`=ތgΠЫೲՀůGL#z11~bJ&.s!d4I;tdZ8D qSԒPD#<]ƨP1hV|U<x:HZ3%SFq˔gt"&AFTxQ<2iT=y^̉M8R,]͠_QLhO'B5Eӊ) We_[ PgtH %ˢ(#Io 61B#sd2<^om!6}YW]=цPzS6?@Ȗ6xӇ6h> Wuf0]\!oiomseM7()۫;q!Ke@IUS>JRЮbv%R zZb{~( ?cR]}Eס"<ۧym^Ѽ۵W^aZIgqn[[FIx0\':JC2 u2vSC*njR-IiT<;PrM)6?|+;5}:5Kl/󸻖fy1X͇nƮz=*q8킻8ǯpM&lG%k=o8Hem44EJ7'ҥ5o";y8\ƗM\/ܫz7muIӉίԮίg7him/:PM -m54)~M>>Ս*wwhb9hMla&-=g}_ ܶ],iH-KݜT mqXnقs?nXmH(hIgw;^\<3h,>xww^7<8vTZ-e*10#O,4l0wH4/V:<3O"'){m$GE;`&Y|50&rY$Y܇`WYe{A{ݒ,֋ej3)vubj]Tut%y1+soOi4>cb\!JMgl=ᯢ|!^>kT6y찡mz'ȃA(k)tAcL 3<c܁Ғ{R>/D V0Sd>r FG3Z%ΕBRieGL66 -<Ԉʞ|P1HL451)sr)C3(@mrր5IfVa8;]Yy7rSa>3u\ZU@TIRʊQ  H“NRMuщ9QtT HG4. /D&.َn< K+]L;EmSMڽqgȈҊ1\@h\ $xkƵKI'4 $,zȐ2YkBDED1 ?8; :*k a1qaCQ;\J?DD eH="nx:QY/T w$H͢^h)r(q1S6]є8r(8PQq5GB)G&|3yaD,&.#C\]uӒ}qEb4 .4vRJ7emdF&:S{gV+˜z\<.viǾx !a[E5r7(8 hw)=ǣjn*(YYi5w *cʩ6p=;˧=JD`Ůje_8 nB}Vo5sk + 4(*y `R:XG6dH`\A;D:B31RHqpx(Cd(\!F1qvN =EeIz|;q __.Cc˞@fXhoq]tSR6ɂ֐-QD<XDV+*Jd4vv M b'M2xddJ 4SܫP-;*DD9N:!kcPq%B`bޓu!$How(Syqql^T42'm Ir3m]#/Mjii"h9זz%5uL ^e4# u6D*;AOxtI-ړFD@׊1VFlmngm4N"Zrb(8 mavTÞ:ס-/uvg]? 8LU}u1ήGQ{ޭiNn&96[tQOf #U ? ( lp6i=XlH)|gnN}+?>ЦkolZr3nMnqѤ[GqѬo+=r0q0̲+svfE([.'{4::FyI?~x2s<I}@, )Ȭ0J(vzCN/sj:^i=4M56%28j=!٘ eq#|!nxg_f%O86#@Em@1?scTnZ4/e8?ߤ1aj{.y0P`uRέ<iE;yfo %]CDfTA2@`x_.rg@X[f7u;4*MH6.YțsƂdQĥ 8-"If( 1! ֬M?_fEZ@8NP\4Y4ǵF\LI%l@x4MT"Ne-{ŀZ0|^zT;Y896r6Na|i6H\JXPqK\cI8n#vK]ڧ+a)!ꛟ|fSH:xg2yH ') R8Og>XDͰ$ LQeְ$!Z$NeIr4QFQSrE N_L* u)i>U6|){r7= `c;>ӶfSvݰ#x*UBS8!e͜۞u^o@(x?lᄤXSiLtAHE=0sZ%g K/۾C_wJ>CʷW-2*{5< QzWlm/Uק()xUۺ`q}n}nzz~l\祖a2oo~u{1ǜt~lO-N{=֮=pްGPϛܮ՝˓U7s07!'/3H歧d~+ ! PLũ`+ 8CTy{~Q~. q\"|B`^N H*!(DBzJMD B,lPm\&[f_70/Qύ|'S)BHqiyىfhGR%6}~5D BJ(\%h1@)* `׋% 0@M$g>A*ɔֱ$U\$P |hTˢ2꽳8%[zW-[s7ZY^αݏݷKjbp6HeRB0<(+?a$ڜ\㠤lxsmݘHcf Hd>|Ct%)PI\Lxau͎ͬ`2ٱdSu5!6ڲ2[15꣞쿵)gin]wmB/KkfGя'!@β4p嘛|: ?6GǓ iψ$M\c&iH;$6)Rm⼨D \$4Y!c'-%P󫃶n~gz;- ?Y:6[WW#Y9q:@m9)2i)!$Dt2AkENs 婎F&r6-G9=rT0T>])qv(wpnǥYItysaȑ6q]޷nz]bWlO נuX,!BAByuhO? )AF \±UUzoG+ f \eqm,-ЮU> \Im=&B)=,.c+V[W(=p 3pLp z<,UVURd#•;WY`q6Ki8m99%\%2_% s|Nx$^~z,Zr3ٺ"_EH߇6qZ kIřJTp Ѭ=۝((s[N5z|AR#?@*J ZH%D4I GQ%ϣIR( Z>d5P d[Fnk6? W77۳@翟<~=;3l._M/?Nl}柭)c\%4"ywit(~k[6%.x)@9Xpe/Fzzďgm?8~ H"NreUYROj<\!o(!XؾZv1AvߎoyQnּ Y0[s>PŰ/Yk9QO  4å3&H}2 Yז{~RzVPe|\YИn.yF2SlixE^`۷uՠuCߎ/#ٿb}͏"`^f0,.vvbAH[-y$%\`[- eII!qlvwabgi<:ޠ W:v{:_k͟oz?Օ׵-2O-2*~wֲۡ6.7Cil+~OXn;^qס3f^ZD[L(c9BAg2D[AuS';ƨ$_Bd3I,$ôEq O3aAVhaΉ*b?ƾ_mOqffYxp;z_NcspDGRhmĩazgw;z 4g5=azO/K&wy|q[Ι+ }7-@Jfd%3;HwDeTL"R m![@]N$ :>YGO@}ihrP׃C]hHI g9n>`:ޏ^XK 6t?\=Fj屽}H]//%+o5 F M\ѥKF$0x vbo*X4"BVbZY*>%=&"b(2IAk)0}Yvc{`y =Hԅ5{ܪ@1Da3D`SJJCI)&iL`b'ٽ*RzpҔ:lRZ YCpٛ\%fKʕd $;UI/U?ۃugg*1ˆ)ɧs_V6)9B|3剘1rm[3f٫_/-MxL.̺T FL# GݹLق(mxOJ(:m5ςPf F1;2Ce5VZ*l P 79ʦ(*anMQNϙ"H%@"EB2u&Z6GugOBiz?gR?s/'G?Q B &XBaRHFZ1i\)d{V@qHk[ FiO}1D9$Q"; *ms m4Vfm՝ݎo}#y8Lk{s@ޚq;y6+ޚgn7LMho&(mf&}Q\}܎jl2[/E)*Q PQ&عʁH6 0Rt̨Q{)چ0a!*1ZKe¤'XT|T̾ɕ3q@aUjͺ*Ͱe1 =6\# *ŋ}:t{o5y/.㕓>$r]>R# =*RhdB=X rSNP^ %tJ b-*OLy-/ |oKE$tYwv#vib j7CQ7Fm?nxCV*#63iDLl 8;*HTS<,bE4J`tF0\82Fi#L3sͺ{R?Qղǁ.gAD`.Z!b p! Nekq9 O1k)b‡ N agZzԄJřH19`Bi%Rb&|Rҩ fԑqq-rͬP\qQ8VepY^]:aT$:t:xfF>G5R|PK}..Yǡx: aK7}?F^4%AeEpC㉪QpoU6}c| lPY$u}IDۡ8eh({l-sWz縻,v8ڿl mۅe.mi}u|]Qq|ߖ(c.F 6H,< G-m$ps>s :~LFd9q{ٜv?[FM'Ŏ߽%;'2#A9/Wu|}Ffvff0@rZ!I RbOdl)dKb+d(98t۹Mac)(CR)A_|\i-ee<@99fٓ=<,zĕܕdhBgӴiݩͷxNJs˦,%,Ӡ> xǡ8I똕2MDdPţNA:amtEʢ w9LBŵ^ج;u wUopsr^d1Zˌ td1EQۜ[b`e;Iel,U`<~[׬<#?ӜB^/JX6Ǥ$G5 ń)hdD,Qro-j5 1zɋkOX@P5'Ұ㱽K Ƅ Ps3Cb49nM=r3Ƃ|pBrbʪ-L &{ahQgXRS_;\M2HJ ;^٠Ybt݅ioJT!$6]@_6"W{kܹ ^i4}?@Rʚ1gcG/0\ hXAZJӊK^2#z!DW(HB9.ƹ B ElՖȭTw8E$%' A"Ei SҪ!YfٝfL։Admv'.{L{NOR#oF~uGKԧǭ*)bx_6S!QERneT|88o]@}˧m7l~Ɏi eoveE3oķC:D/[k+ow待xLZ3!6t45)xv3"\&{Q=%_JXM!*jl,I[lht"7gC>"IaվtN{snG3LJF፭/7/07DǗ-tT%Q*̄1A"e * :&Ֆ_adl980^PTΤd}y h#?$Қ$Ƒo#Άl/3nWnb{3|܆nyI ^,81GU E M9nk`VfK?|3Mf63oi3 Ϗ<2*Z̑OP)S Hjܟ}a|* rJ碜L:(i<5\͋r^򉎆%XS`l\NB fBF9&; i(igw7h8`hWfv|a}dxW^^ >&S0lvEk6#6ECK YY?wzBpay̯_LO7iJׁ&w-^mf]e2uǢ`V KH[jPoz5vc0Ȕj]SEK'տ~sL'W7K?~ϷP=z$=>mݢGz2ښeWɗpyuxGu4_k}n[7Rz_73m| M¡M M__dž->rO/چ=^5.0њ}+–75cCˈ6d_6T 7{~mmdٯwF=U幭0 D]m_yzu__y+][hcuWtWRUtTdZ.go[)Ȗ;^bHI)WR̮wiU]}ZT}Jif><{QrRηJzyq+Y*mK})~}-/fƁ푻$Ȗ-ջ[Ƴm>xO2 f ܥj /hyn!)_C9:f\߅o[n"Mv&Z֨~DƐ94wpf^zf@EFel&qhz|HM$@څ CI*}v^5CJsP?&RȨ5MTHڑ Q/&#`=#PY!B5jt>.)' YE/ [V mJ11028VIx'㐄of|Kwgrf*d2yόBaXt !aa(y0s$G6n|{?.4Ѧ&bAmgݫQMX'12hkiV9m`'1A$59$! #}$Kah >=mpODSg?8R[G:$ ʬ_hRZǓRfFB*+m#ItyD^؝`̇ F׼LRE*R")xH.,V*2"Ed ceYK8|:h (b0([r~`;u8;?燡C,Y_z8HŬ=l-oq[Cm&qEN?Z/xQef5w\Kʧҳ$ᤊ+DJ-?nsWx`sU*e9d\Fʼ?QB.2Yy3nPgD*n18J4cΡˍM?xZETO>GNYHaPNՄ\])Wv6C7v52S?$BQ Fb F` yU2 PacPbm)j%i-r٢|i.~[wI|0bQ-KRHˁi"uj*_L=וe_w+2AC;Y(6oq=/np`IYF7KKy 4n=-'{.؄"qrq\(c)%χ8"y˴HYƕt5Rq5ɩAb.J~n';<%BYEME85E7"e"PsW $JDKUKm%zQIĥ}FFT_G9A紁] hxAÂS.wR.̥W˽xV#<_#oy ;@3K F⦅\odX g+HAW:w''s8C")NQ.*nU`xB}:I,2{OFscc I2& O5Zks 2}ֺۗ[Btr<jցx.rS=ӕZ,y<3yj\j~vV?w ڀ9 ^ZP0h -45.v0-%/U[%՗d$[20Bټ?/4& 3:omR$vhAb (w[ %Q;b5Il֝=w**ϧ}֊TNsɷp:=8[ZYʒٹ\3#zqzɼgMŲiY֮Uwf 0'=n_2κ sɭuT>4G$2;mGwo+-tX}uǍ*e}VF%d)/|^z_.U77w/ŮCW=7?o.u+h>b.B.ω..&eF2 ,Z#ՁHOhksh/&o!t!S?}[)Dk4t/w#tI$g/1#xq;t&i&(\NvmPW}JFu@ BVUNk~K绞wU@d|)tFOEEI,, E%0E%Gb.E%䢒pQmJ8ML\NlDJب"'IK !)$'ZE&Y}ki"uAgX~+(-`O#1HZFug7+8.qe^KJ)(>2Qm;ѻ^VF!o,ZW Ax!Yzu\\uHM*|Ӧ%LMuT %=iu U 7V -" i hH=D6rKq* ?R*9*%QZolǛ\>z=GW&4gϦd2o7 A'z[o(+ gB~ 9sā#,F@׃W+ӱhW' mmMRZӶ*BR3)s&}-bE[y7ގp_oGYbXKWގG&1giA*Vju5 c~FS+N^&8қ}O&[\,25Wn8W(뿾N2sWE16"o<P3 iM+ t2HH,*:=6MRxMPz"{0J@fVD̉_l<7PWTZ8A:|I/!ϘA KH!)pY09Aɱ~@wrz3AK >{:zg|ZmD8tyf"ʹo'p4b}n~-T-=GjWD.;\ryL~vVYAfWs~i450o/]V__S6^X|`0mw~z)_+[ռ p\YhŶ\>`h)?(ve% ! !Һ pQA@.*yA[Dl֭Q,<#p% odJ@DTԂk̥wR$i KY@\nh0#pcuP%*\'>Zh$"{rjp7Lg޹8vLE𣡣'#ݡWz,X/>ΐ^B$F J=>X\y<XC#w&@)Q P)1ZHT; PPyM:AF#|߫%|xpR_k[Ex{;Rk AMak7p[xI9E*XOYStL1AB:BՎPB U!xͩ 10 T2뭳&(dHƣ'FT!DT6Q t|Va沚L_N<v9AEqz󡰷^:G`sNGfwQ i a$"xICt(7gyR i42K%Yj#jk%NԽb$Tɯ3kugoYϥ!WLJfWt+-5,%$M%(ج.`4p(JHt/3K&eчTU#Ɛ Di;t ::mm;OUr|SKRMi$fS8S\F hXQN|fCۚJCY297LW4Ǐ1GoϭhAR82:("PPr#U{z=T]LRIRnN" zȦցOspR4dl\zn6|~Kx ̝(0~])%|/ zqf.7*.7/qMm4gg8?sps.ouFWGa2#}- ;Ã^ ;߰v<[ŒH* ՃIygI冱D Ǚ-$%LRaƧ9A MQѨыN& D'>EZ[j-G~ܒh9l[,#45+O8cӥؚ)+ I(;|t^8<" 6KXy71Ny[QHGS#5T]mrڰ?1M>t6ο9Sa]jJ9Gim֒t% i h:=R.H!Ճc(KzdJZzA8%Od Np n"ku(7i}@\. *A~ n; DFa=!.9"jg6N[ R)OR8G]̺OFr/tEb_3?|v3/EtׇԷiHc*bZz'BO>Ȏdy^rwlyt%;s O<87a<]eV1ȞΨ }e hbbۮB'aǃVd`Nݍ5_Qcocػ9vf=pf/s{96pFܞ%$2Qσ '9Ke+u@ T'5TJogHgYwRN B$xDP*A>J Gz!jx: Øi0zU^6-n%g58* uAq۪w\Hh(uyJ_' o+q~Aڻ ݾ\h<.vmί4Q76{ݨdcbc2p-2 4Рi)-hZC{X'Z𘓈IylҤѐ`g.BL`dLO*t ud!%YXf.% ˀчR s``Ve[m8 4pdوho4p.Dd}6ܣN7TAH^a*Lª9VUԤ8&pC]L:[Z3?L^DN%ZZ 1++jWԥ7M9 y> K&'RNӷ?tԊY'$$?D{˽l]OnWe?\*PJF{Y!j!HʝLB%l)DPBA  tՅ0BF!0=-3>2=)h5`vAOhsVĒD**ֶՆ2*la58ʶ4>ls.ovm\L/٠d.s͜Cw\LWX# qiV2&')2+z yL (fSFCSN2P.)grLE-va65y.Vtk]e= ؝''*-"CB -QV(묋0PGpEuÝ&\C&d(DI$a!H6.xNM!RpÎԯ,a%.C^)u2v@| v3:O0 @(g\h_OkHD gHocBVoosJ!o|ș@w,en:gJHd^rITrebd.4qrז J~1?ei0 0Ih2 !)P^| iDH]4ٟ,  Ua?r l Mt9\UԚl*)O"K1Dǥa%}mBV8,Mji:HPOdyRZs]Ѧ9E4c F TW!$2.z“'Ȃ s`q#>1fcN>#">KD~P;6qHA=(`L]ڲ5e/qԦInմlv=mW:On.c<G/4}ǿu;5Lpr)e KǑSz.7˝)VM1c4wxyE, /'f/Tr ,,Y%n].U$=6zFJ:x}J{?Mղug[el:$Q1?Mz67E^E)Lxd1kXĬNq-")mM<\ ZQn.5OkXD.pK^@Cd`ՂuۯE%t}n>>&#ID|"<]׺x|KbF,R)_5W[K#ܫ[vc`?ʄ&=Z&񷣮 ow7% dLD}E\*(ak!ݻA_vBN[؄opyiD2)+c,Ux!y/Fx!p[+lAjn:93 D)3.!Y=qk9HKK%IpMXLeQ6sR쥞Uj2q('};3w~<81@ypx#+ABuGY-Em'^蛫T_Wp@^Egalg >eo0W#K];=}UBԯY̦2[1y,YHKC~HPrN2/ /sAXG|BˊQY sD~WyZ:6pd9)#r5!S2 "Fo i?sXPg,\4 УPayZ2/5+N则'3S< '4UNhB7, g@c鷆d Ke挗co1CU/Wv⧩ H&j@$Ӕ,Ys1?>Us/~. PyeGBF@> C9 4Yl5U2pj{3!s<4.UiO7W罸o?I"FEAk!G2EZPIZPIYR!T&ں)3(s }I So3 d]V +NfW~%~D]}Qe.ڱ)x[ML﷜~LU ]N{>,Q= +IêsΗL?v1XDߨcTM%i@ۘ4|OEQٙ(*; Stq9Rc(|qÔq>YJl94[JǗ1,qvW4N+,vZ9 E.+@bZF<ݾLo d [Ӊ x&^U~Q;W{e,%a2ݎ~-M_:mըmC奇#?z =mh{@Cm~:m=n|?QzJq+[fRi_46Dm2jcXh%(ôW}m0]$-ς;@;𙗌5}[}iAv7ن z)m8g]=S >pI T";&)V5BX)4N-E\U./)3dxL*غK%-_12:9$֩3 ]X fe J7$ݡ9zWzuc6K@+{}[qZb \*FY>@BFYt"0:Qc6x& H*2O-k,#RN}{p~E/ſn'Dɬ8h%WZ̆O4=vFAw{ci:俯} n{]}x偻8^>#}g>qQyRd2oK% ݃|=oqۓB6.z'oOz;!w^.K ђ191v|옄Y~I~m} ۂ7J̼ov0›bdV% K\-[DFx4Vȼ›h|[;h;ww߇z_Udoec{׎~|xw])_q`GÎ/لjnu761Ɔ? 1#TddS芲*e/DKlR_ և7Gy` )k1Zp&hL t hxN") B:C 6%;0BS5 T&;gOuk|6);`;Q=Z囥[or݂ѥu,Fv6Fe +5`;h4Q]W\WmzT.7b(%r$JZDJ+`:c0&b 1.I@UE̠83eڤfKlZ?t>Co딳Tk=1 j-c6JQd}QHFب S68 IF`!B.BɺL &K.ۜ]$ƁN Ӏ/#s_ԭ3|u@MemU_YGp-M99 E lє;s| p2 .xlzzg@ ydX'LB+P_p z(/N)'m %P5!9eFk0G0{&sj3g-C%C:*RZ/~x' ^'梯ݢ&r[~0=hYCKbh|dC`dF'KpGxڛs|^έwbmrdQtyx_f6/=gK,/j>4_\Яkw"ƃE(1/g/ի?ӧ0$7 J1SPޢy?MAz85\ zt9^^ݦQ\{?Z56iJ x8ڳG/DSS?y_QƧr|dSﬦ֯p2d ?uRFR,F'qwB_]N u>?^/y\\լ=u310!}|_KͿ=trq>m0 &zW'F|Fvhv,wΗM{|~^v2\x.Kijxkj&ź,b&r=_v<^x:^Ș J2i *lPǬ= ĘR>AX"@ U [ѯWul<3vOоPYEyƀrlƼ'lQ*ў4JS4eۧŹ?]unTu[̯j4-QCRvDJع*^Wiήl>,Fd.2'31u"U]Xi#M"Eyᓗ:a)i I XR :0jLxu}N6}qliiÏ&)@ Q@ɈVXN&_3W)Z)m?PB&[Qt6e Q$J<ȎT2@Nu"2QI)R DJCZ0RLBX(k*ʭYw#c; IƱXc4>ʌ{̜듫{f_y叿n2~*o"2c@>'F .dSHAkL]u1Q Ҫ¦vI L>eJ"2ƈݬ;;ͦD<n&Ǣvha@YLJio&Ì-&B [>c2>`Sۆ;bNJ23CiXΚ "c"Na .۬;p{YT`/Zg1T{T$tHEX9`D T3ZW y32"4uKY[,DR)8٤#L[rH.16f9P{x@ӳ z'jOƬ͐bv;ԣXܟos九o>hˠVj%%,%I1t13G) l\'Rm 9@LuO:G!$aA!v s) }0:&Yw^;AV#-L0Yi(,ji<1*#GuH9Q$Gc!KabZ c 44D)&eR6RVY+A]D09G@ %$HrE VUnffUD8 Q^Dzq0N9iT sx9O<Ώ{G 'bԬj,LBD !#:b/:Qc65 6I` OCcϫ(ixi0n߯ & 8hui~ys|7tv7/ؒtn?=R2lsR:PwTXy:JlP.^<}Kj@Xi|g(.i;'E@J=;js0 YČRde@@p-$md!\ʖQF %0d"gLWK90>:#ZYJiNjp>f.L/`rAjʄPX6sN1eJ٦YQH`$\[Ɂdn9oȹyFHOV] ~ޫӁ"Ռa':PJ5Ovyŧ凳&DŽ'3Cɠ THgƖZb۞4ibnII|_}4m-V/Q>MS $bLvUʀ:_C_9d&fH%e"j`5uMn5k}~tCl1Ev-Niy:?EFJNɔ:a:c]w0T'uQч!<}o-ч!栺$Ӕ1/CEd9U#HPJٗ~.M [gXo - : O"ږbɳZ`A:mPVKb_6^Od<(;[6_4!i&+a8RaW`2Mͤks.IA'9m kgѹu.iZ{>r?=RѠ6۔I9#cRj6J(#0ў _C>dĦOR #hƭu*k<+K2I&緟cRv~~}3Cǿ}ߨ֜r-T'mʪw\c JIw<o90sι.6B7^eTڗU3;Pôlt%5vRii2:ͪ$4IWy>]x[3t!}}Qٛd>ps,BkrSem TsjY.(Kr9'<4Ï"ʻ4Q$93z!ɻT:mvr%#bIn«^LY#^V?2vUkvVtWnZ4L{׺z\\ܦ1ᴨFs]y3*7mHc,&uˉ׭؋5]bU]B;qI ~SnQ:uuBT![as>x}}rR:XR) WZ2 Y1ǸVլJsj ;AOncj es h շ2"b$Dh@D4J&9OSLNXwVX:e Y \==rI9;ɺO/̣Njf #γdU ^{_]AϯrϯYy61x+Mr!3\I(~"KвMǥ/S.E@fAUY lFc@5omzcx[wցShluoۻkM7n$; O3AR>ԝk>a0G:SuЄB\[xڔRy#lȏb6vGؤ"IKʄ'ܹr -U&4p`:{mTA[Zb<VTM>š5ɷ_G c^js ꢺA[H4lE}՞ q ~r$\3 3l;Jm{x7Kݻ1jzyiN(T_^Fm=/OeNHGDEE :3|@άiv3 ~ߪkN S۵q~peg%P:o/XV7bklɳRX'5YK?T_Tc)̪C>{gCZ Y ]OȶEjCA!ROir[^rO+!cY.oj-j5ĦHٰ]={7}9l-H_l)Jn5`εզ7#ی?S7T  odpBbP.;ޱT.ѽ R~}dgD ^ m39H5UacK gc`Bz;D"eaM`]^tGRfL,js#ކ{)mΏǥBSC=8o>z:?xzu|/(ӥ48̵3⋦ ͖=xR'KD`?QǨ}D%:E贑(oSUPFi_to}t $} R3 D 4S ZGQ_1&)N/r.zrHaVl(D'?6EdzI5xZsIŰ&i/y<$^-T)@Yjen[8GpyIj1¿(Y7s/)\5HN3ŝps2#}Dg+LLw&]m6^Zk) 40b^̜GlW\Jj}8IauHl'L!Y1[ ۑcE}f?A YS;û&Hb~yJ&E0_P?;qo\M]K?~] .^T 9͝"1Ro}-eEβ7Ū.ڒgOu'הųOϛ^m-lyr6$}^E݆%abJ#mhii3'I`8l[?vPKZp(J;ev_Jn8hvժ25K7 lw[+^Z"좜_Ry;cG;e!Mk~l87gf͗)sA!'}d*242!\6mL>OR|J?w=׃QL Xv4biVъdנ[h⭓Em^yI\Pi4JkbkM2@&!ѳ^S9qpEK,D2GÈ_\/ ;jv1J1HIvfM5) 2g"P-o%sx="ŀ1; k(aF>ӭԂ[š\^vRRW:KB)",H{ )2Q!Z$TC]Lti QI3ĘS eP>TH26Q:jU> {4Bpec4j(4B}cWp2kL qhg  y ZCaU9ج cV2rQc*h! {NF%Visiv9t!]21ڸ@Y#AQ gm=KZ2 T ?:RI-V&gXts$!G#M2d䲥1B$:l_2jM0X*3J۰u gGXGgf qd>%x5 "\hK e8IJ6|ђ(v {^b!p#H F2Qhg7f6PuczNb +g)!0DGmw":&DiFܶ0IPFFx D廈5R(+)$VhBv (& :-VX vj9()v##YYԽN%PtU*ba%^fd0{NCAa[pE, zZkHRpj,GGH*/dFV9BPX"5R: cޓt@,@ FKJ8"F.;.b/5Ob3Z YFNx2h*X.I*L0k٠* MHpq7PʠObȠ>hHoe*b"| ǒNb)7yHrIa;(Հސw|Dd4 ao DbH$[ Q!&K8!N &$t8dRP 3)J;d d7]) ~#XFd6S TIVR HȽE4i*%!@qp0seĸ"*h+gpZ 0?#|!o1AAOGBMvkmHBT+babk`H I@,^DI,I-ZRwmȪȈsdEbk ]ZJx5=*(I}^R6=LbgV J Dl?P()46 Ja<| $:xeH WE7 JΘ @<n~[vw#e[Ŭ*$c QIbYjˉs!u@0- |~n&>Evl4-ܡ*MNXXV˚zޅHP6yD-$L>nx #. f@ܪIKD@KJ7AyA3 vXmAԨƂBG1 $bDN W$2+Bi,h8=Dh":t@2ěq+2pf,XN~d} yrx*֝fE 8-`9%^V0N+<]֎/槧,»!O‚.&jHmEZ6F84=bTnDl \*&@QR"d2[{G:&oCUQXÎn< L>\;djy=*f @cb- FyN ЂH+Ghإ6FuF 39(kh ` e@Ao 2ZdpX0FaE^Y` E*C DYl9i L[B:, f,yՠ6*gUn2+{уEXY oT[ 3]$y͸t,2 c%`&k1Ҙ{Wwv9NiW>]Q1[}`W@7]pD3 L[ n+~)8w`)jY:F nkH=k ]5jh9q=)ǧw~5<#|QaA]ျaQ^"o* 9pyX%\D[9DKh,wzTaH>"Kk=xP@z[`}fFu&iD _a+@1\ƒr\ \1DNa}Q&xyƴ2apaBlO(h(cx)U,JG52⓪`ag b"Z2=BmX)+̣g).XZ#eʍL #tԬT gF (I<(0SLGjJҸh µ;k$GjyJMg Tf]L!wxX@a ,gfFXV2Zc@+hυX4S"p#FD>8=̠!FX6 nqL*p+7^0@\0 懍9mFِCL*V՗Ko& 120+&h&K`$1D%N q9,Xc.Q銄ީED8Øeijп7u+3@хx(SdY/[>}K ƮR;` BH~ V߂VKX],^r~wZa( /eѶZq뷾\va؁%\[^o?dfE/5r[1N ;nikX|~[R[}~su};e?};Ǔ_tZ/o䨝?7/V?qlgZ.O,dӂ4'4`l粚|Sl5[Eu߄;ԎꎹN߸=X?!^ѭW0xo/{Gy+ms/^p+!Տ(n$" HpE+\W$" HpE+\W$" HpE+\W$" HpE+\W$" HpE+\W$" HpE+\WO!sV5 \|=+'.qI@c(!֏,ΦE\g}KNgc8^ANsc_!}<) k#2BnOJ~;\5nr9~hslq?{h`پ\t`0]]ACcj;>fL>=]/οpzpՁDWɖ3!-OJt^SUm*0sRLwŶ;_E#Pex\!zT-#(ȠIpRL)-NKǝ0*i]e*8,2pSA9?jQ`3sصcgjTqIp={8xޕ״ǝȌ>FEExVGPw"âH@`O htBIUMcr8Cz PڻnTTkVUMʸʚ5N%}Ea(6icC3\o[րxwK)~mn>»G'LPWގB:J͇kozO;7ƫ6JR9`ՕQ< $l!r[Ttz&<NuGI ĎcyKQD ؜|<9w8%cuWuyծk޵:)zy fLگ,t뷍>QJsm[E%sMAʲ*-qK[ |A|KǷƝj hsL W,%UvQ2fS֦lj«60"xd\aT9 cFsO /9#D%SyƝ_1+7<]܏޼D:|vhw02OK+YHFaX m(Lb3+ .16Bqf.~=-XZ7a۸^e)h)z^?IMj^Ý h|uᶛug} t;z8XZ] Qa~WixUeaxϪ7ۏ{Vy};X=ދĞA?ᣫf/Y>Oi_oG!E>A}HR䏧L?J >ӣbカOLOGRj%KظN (E gיWe{^Ʀ$u53Y.U&.ڹ\UiU6nqVrVTynۄ)2 ^bVi)nc6kUrj/T<$1rm4s!&.Z!{h}8 pkc\ ޳9oaփ]p윽y_>-K/Zl}mRg͔bj>b L-B%_;ݠ6#v񘊂P t[7o@\2x{W:?Gֳ1͈IPtSB+8՝JʬJA%I*SPޮ˦)?5݃mb3)ek(SMQj梳93KoG& F 'àv}Nt <_mw{Gi/Nn7;I}z?o!Z霿_6{tضPkʩ,+ ̶h8i2SEt֖l52&)";n-ݺ|ǘVibZkyDz8mL)l( *FT)xMTV3hMPb"DS&/(&b4Cb}@)y2\wazqPA+kop@i1CpiBAK+f#jbRg$>[dPQ8d0OE-O 6,J(,}]5_ 5(Ǎ\i%#*FW^six t !*RRj5~Lryzrs`f pZ\n\O0.g([=Do.G4¼;&{A\&CjAd ",{h4)91pJU]Lg샴U=Tߑ*ͧhqs/]f1!ݸIw=LbO/8g?YZ9NxKsuXi4!1bU q8t)8KKrҹ䰩#[![5笱05zu骣厤}Gᮻ}#_-u QUrŘGNf>Oj|P߭-"X޽x94O'=4Av2ψe\ iJ%M?) u o";ą^Nsy:R]~3AnQp 4nce6dI[</Awsbzs젨7F맵,w&͞>9H 3~&oxbdTdj [gUz=ax=54mٸZ.>wn{ͱ{73]4 BB;YwdQyD Ɓ`l733T7x;)Ԍ%sD5M$1ZIj,jB*@q WDe n`q7;|x|O>?Qągp! CSl?i~XiEwQ o_vj]7nM.hS7 Mȵs&9Y 6cqŸi~AӜ[4S(ZCtѐ[%Z'!WRQe>6Q%96r Jed5P21C7D/(18-(CNH]rgWaVa[Cws8lDy7 i Ss]0"%E]Ti'&gȯDLBfj *-Sl . +cF($qɍTyC 9['y>OqmqùX\ԣ-m"-b}ŠVj*_ 4:=:HtJDL. &t!a=0cp01ч F꘬gA4) !"h@IaU(_3*ta18Vº 4+]F[9ibfCfn<~W@Uq6 c+zb(7^9ãʦ'JGs&ʞpÓא=3βJ%Q3LYNMLXVm kb~:k&hbqֶ`w,zhFHCLj- c J-b ѕ wJ-dDT8h501`&(CHՠ+_F"+Ƴ9y.GjDNjDN8h]/P'yTRaI O2 ҀZ`$[65  `j&ўNS#7SQ/6Ks,%EZX/A/zq;ROsED/TJT猍pg2Ebd럭SA89B)bq>d!;N@ 5rn8NExJޏxq97$tRO,nҼ?j@3!&Ӹ=U?=9R^j#, k?Yu,|Zrn3ܬdg(+?-k(h\(Gu0sBr+8Lʈ[&!qT68K˚䍳s}d.@;0d  .ӏ5WqDB*4`5 1RApn7`Nٕ`t$}'EOTQl5K9Go`V<%C.imWAUł>|0)EwKäPHBwϹIqyi;9fXiӳ Z镣̆kDӈol1'oXoSS:9>d/ 84!x>+ {ӹG I1ǸVsHҜh1r[9Zg ) MѸ1Yʈ+ E,a"QmQ=lJ=x(g]Wu~{qo :??=N_ܠMˊ/'e_fڍD83Z:1;jN/x쎹vL{1Y<h+MR!1\I(!OZ @h> ,qi#˔!p[$)="lk*lFcT5 .ׄ싑mjsJyd!x_+W&__q+\["8@ׁ>PAɉ3̑NrHHv-NEMXih84o3;;Ц؉<"9/D-Zj73"=]]OHNҚix~_=s4H1˪uݵ.X2[j{z^ﳞkyn~O;[_1~xrz,zkqFbM= ^8ТUP7NpWlsjm8A͙'$;SKCH+!*1j[ 6]84 `/ڟ!ۢ-.21%>1& R&4O09A ]?_83ȵiJ4};//1_ brcly岰s0Yn6jj3wK䎠,qVԚs5 @!H)j*ݹg(?j % @B'/ҊigI+=|rxHr5(=Y"W&DWRQ\Vp~Xsa\&wQ1#v@z\ (:x4&qLٍ2KB"FFYCs&Q8S96څJ1p q1r uaG= 9KLm6s2r17-ߴu㿭S5T[۵]م0ts-F. _U&O?:I?ME t/e>{:1q'2w}ЭJ O |6N7.ÿиc K}-H/D B:Wן0 L[R[~&m:ϙX*'m~ ?=KVljH*T-/.ۭ5Z+?~UwK;eυ[uONدCuirƉ`G?Q}V"ھG=0 cן?4x_'j)` s,N~kmHʀ_ξH~0pwI|acxHȋ߯zfCrF-ӬGU롑:26)p$~v.,k9@+dnkTc.]Mo uي c[2)np09sH[ *z]ʠZeѭQ65XYj:Rjsf4H1Ov[BB,mr ]6)8a[L̻hrVX[aPi7gUz٫7o`_^dij&7l݉?|ۨ"=.J(TLGF f1+rhϲM*]d#Ͳ{~W e?Ȑ4!  *aN GUQ--JZЂ^f .ZWYFILhIm VLbRŢ*'_0SgZh1TxeXL_eK="igهj^uidnQ Z3:kAYCҾV-Ǎ5p' aK<@y"S-E~&ޣVKѭK1neOč7<-b+zY ?CnE.0TIJ\$QZa%E{%EGA a .#z~\arTh ::^n3RoQHV'.8Ha\Lr\aL|o/^]9/-w_u<̍A9#)<Q l\Л09g[l) PS?FO?,ܦ G:.+ Wz&z7z53kZexxa6,i#dNc; 1erQRQMSx|{>d W#V360W9:1webyxw!]ab؅v!]abst|dS܅v 1B .İ 1B .İ 1B .mW.Ja[zZSM`ksC6Jƀacdz;xVܢfZq{VjKB[k㩚 K蜏$ q48o [IP>gT{/mWڟ5%O- B9F54 2EAT:/2#0 #6JbA`ѨR]駮1]p$\a}tMX?E^8[6vx8 "[1ӝRFֱښ Ř2Iα:DB`1 N(jϩ 0u'H ^uֱE96v$4G)IFxXExEnN8oXi^0B*E,N<}(܊i:s oi|ڿCIb߆?#\܍gDYV.Oam8(C%5&*Ad7s+g1E$acH,>.ɜG7:Qْ!?GũWǬ8H,\ ؠȨ7/u$a:Lg/)|p1LUN۬qS~c̮F0XCZm{3EzCirggXeNSDlN;3i v0yl[  1N[lB\:$U]NʀТwuD2 Z[qdAA;SmaU`0 *Z=ؽ'%<(hŽ iĭP`2T BdB*$'!GA RXh!)Vt1| zV%nϣTX0ڠp4ƕ{رt`X1XE=ewQ4no( 5x{v1]ɛ9P==N Ơ n˅~[>_ޞfkCFp0EN93X[exv}^tЍ#k߁F5 y!QEB nߒINwa!$2.M-tJNh?hQ;ϧտwԬ|_8mkgs,`M>}眶}FWqs4D;pθMop9].‰P, y- $&xO!q,s!ĀLh^8ZE2% GABaj3jX$"﵌F-MVHKD4<-u=a8?5~nC3h|0+GFՄϬ=o@pl;to V]9xVЄYYNt>/vNT?t0^=D9.ƟZea:.<3̠0Mp[BuJ=[ Z"PLhHhÌwy-oHRN ZD ,fxf!% r4l֝J/3 gWVƳᰳx7XV(/mۙ`j0Uf6qb:5QOݠJfqo}n2-04ҟ]e7bK"g-c}hAguwH͗g?P^4A0ɵ r B3f4j|sWlTk%ixkκm +as] +s\Yhy(>r ڥ )gJOf߾-}6/m9.]!Y֠lݰy˽"IO1!t(]<_{ P]KÒUDsw=ԣcԾېqW յqz_0^2, 'U1w\X" wLunE[щ#wb' Âg^R"|< 4pru,2c^fC Ia͡Ym3a|6!D=-)l`V :*B0Fk#40lAA) ę[`1VAh 4wFm6*GTLHԀPE` n;KVuJroRh&,XZv9l)J/z,ɲdprK%6*o`z-%\!2*Cz Z{1[,} pZwB }BI`09vLf!gBZBsId+[d3.r@[XT 8Sy4LKmT?UC8ejۚS愈GȆ\l(n:z&tYڢo4|ݶB䌂~#"_As߁=/8Ld98Zǽ*zGs,V+e!H%Up筌 :q2 (jjB9oDo-l|ԘfJ>Ow: f80#sXVKGyAĻbK!Nx[ }7ɠFS 5EW<] ?2xYCkV|7cPndRAn#&:JMd8Ơ΀%υfz ]?<9Cn,5z[c9hߴ:FdT™&"#lt &rfmHXLO?z)TVA| v7-ӝTc֦/%#:q^`6]LG]AIRy( 2 2x1ihwS ϠZ5*pa-3:T =@e;'QBc#&$w 5|>OlZ) yw 2IVtx8Ys}R.Os4bR t\yo҈<T! H <h8F8)vh2U=46Gtc &H&#a+BC$ ;ja;ӊIpݸAˍNSϬ9^CY !%3%WV+u !‚ÃHh u菉5_wB_#ТFf6CCuQmˮo~LNPIFI7#pOb >Q4U 6NI K|w^\a ) Hxf@RD*)|dt,H8j!\e=(y[H87F.G+幛saѠ,̽nfO8\&/hl4K^-5 Wa:C\2"8;+/x|g/a[S M_gq2.{FKvi6PJ-te~Pz``@ax(뙅T5H3 ף[3!O'/r O—"Ee>f?o_;L}I\4#{:j -esS!YH/޳ VDKV=#DP!uE(9ܕQ W}אh%V_E8]f K<2x.-a@tqTimJ0Wm̟^d\~^@ٻn$W?%6/[6dE"me#oҲcey6t(^>~U,V-c*]A6? ͎۱VV2 @zqTDSr:>Ց\M6n9Ύ ^n\] ,8rE^xX31ʅR{/tY9YŅX?jME1Y3^p!4sU)&FgÕZEUBB'W^P{yz%xI^:i&hGI/L Ow*!&%VYXf.% eCp qA\f>w*hC42֝=٧lo1,%s"{kB[3WdIXۺtG\dʖLɋ Q)R (*k3㊺Tii9OӒ˩XTԽSͧ_'.(Ym.'$$+ ǝM>zֶZгhwC?-d#wD%9r'uʱ'2# qVhk2bYi#IF f6g&!Q,IBbmd֝ǑW$c_,XxHu 07a{nveݷOƏFϣl9,w.`Oz&qiV2e$S6R@SiebyL ()4!)'uYKt)hleĮ֝;LƸab jW}QUFmףv` fBZP216g0$F"aKO κ53pDXD_FhVd,`r +:dk0aIب8FSQGȨN׮];akԯ, 0s-%zDC:KfDQgm#lCmE:*1ӊ$g!% u3 mHC\  ldI=)EPug:.V&%"xsrR0(227%";\ P2Fg]oʹ{\<. V}PTCz_1+z\Q;я/hRXяY%`F&flT{n\ހq.(ltxGoO."$Uip',exГ x I6c${zzsk5w+0l^3u_5k1:, L [.P 椕N`HCB2.ࢷXC#+`-mo¼ ceM5w=hL,ˊƲЙLؠNL)dm N~Ӟ;"ϗ; ޳/Vg5cTMp)01*,F6J:GRl<#rMDC e )q!J04q uz&UҘr6@gt kugGa'sj+{JOdMf~6aX?Fm?{*klzC.@%e&2905.,Qq< ! Q 6A 2zxŒ:OcZ9RQ'ТQE!wu Үp32I"l"E 1`N8$'уׄF {e9Z b R/:3, D)Dl hA8PĆC_W` Mfzr@ FqT8DTXZ6;rwl61B9'gsE@ZYrJiOGΡQ6GٔV,1!}3-/1px^#1dRI%P3#D,Щ Qlxb]Hm[mzw5v뙢nFzHѤ>>_`N}3?  n&9kk^oqGΞvr*LQ5M[ )[,Tڲ:|p+)Pr5Gm0F<`&;qQL2Jsڬy0q)sqz{!MFi5e 2*5Q_6oߝuמ4ߝ>>x&?dgnhep׽#?D3|^<ٻ{:xM춻}iZUl~gEǏ{z;П$`1X~|GZ5~l.f}}r(+ &,qUzk՟l=,v۝chʎ676ij'0--;ӢO~>Q6@4>h׈[@&emT 0V%C#!lk^7!Zn 5UЈ2NQpMYxΜ!S|GZ6iQسbGÆ36;8`9XϒޔO`zwT!OOr6E"zeL'&霹,іsȌ''Ľ'T7ޓn}8̱01oߕ nͭZN&a Q=gaW>^60;/'/HeE"^3N76D0BEs w)ɥ >TbP/aNB5b+PLyZwhИ]$-ς+rCcy]'pwjd}>BڹϯONv;ǝp+֋Ov3V G|6A";&)VD!H#.028IDVω,ANs>I69.CQ؜< 'ʑծX;kVuJzd2]^~?k.:ݦXT+ze,KB{qOܧnS7}>mO+KOܧnS7}>us{ǜ0F>usѧnS7W}>usOTQ[r_ T djO3"6+HZww&N'-D?m )c] 篡ۿsǰu1Be13QˉO3%A@t)^y(r,M{r]7_y䜥q޲t޾y!}KBC]2N9;8P>OXw_7mVG@;$:Q:cP2I3`;~tJTGW1zzᐤv*YwzW' eM^"sؔL9b&Y&)9˼\*^$)/Ih2DEbC / Qbf2xI9Pa2 VŨq* QѢiUi〡G@ƕleI?͈UTϢ:b7nf=m?^ĄH˄/iVȔQCby O>_܉_!p0r~CGBJ*B j B0bR68+K1&YBR2sT,z'xe󄔾UԢCgw5C!CuIe7 w^,NVK5@(MЃf6q{:s{iONdNj8s4h{Nn?'y2JymNDx0+Qd- ~f4F( iʣTG'mQ寏O|ftr5/o»8*./ k pkCc-"{c4422i=Η D3-yfޭp>?vE?$`Q;$طIѤһ5tt4ٻHnWt %>/rc$rI z Y$h$G+@{Yj;֒9fXv>?I7 s,0٣)EM.HYGDW:RM:*wz0;G^׷dNɅǜ\/O}#<(|ӏ볛Z^z%\ q6W/A$;ٸ.Ȧ+9{>o)~E0l8(c 1S GB*rc<4R3~ jMο1a/&\4o,sL%J`oå :Ë#k:ƒͦ {Ns%i9[ d.Zƪ/;<{K {\:K ˿ }5>fңYU{=]!;ص}6 %__&i9I:>F;Co0s^tw92Hca?ZSFn_WpMĮ5:UЭM xC5+>t/4i]噂,;FE\x8[!J]gINI0( Jtx1aGjm_ 4sU)&FgÕZpJfÐزaHy<~v^ s1&E3d\EX٣F>“JOI,,r2`!8 d.3rw;!UVgdzM1YЦ>r3 RDd&Gŵ*%c\KH_fm&eU[nN8:pM3 T3 E 4IĩpT9l4; B:+jlFYx4;ϵcO(Y](",&Y2#:kf 0V$.: IrRbnjg9cq@rGqD.v&`Pjnd1&>)EvWg.W:]lǴYMK]'x4z^JET)sSRr+#fR:.>"xk$WΥ|biǾPTb?{xPTЫ[W6;z(`NяoH0 ~4?ts;jFMGyYՍ \6E>=&Qݣ׌{G%^J%2ػ/Ńuٞ .y+`']Z\7Ѕi˽լYzJ6#:nzΆ K-ɣ3X*MH" 8iS>琐e\k˖䝫k}w`2rw& О(QXU0SsE/ِpOA BTJ (Mjȁ[{qO5)A=9մ,f1cj -s-`b+q2E˃FQE . C9 UWQ*+1ܧą`(Ћ6H$3z1r_M9̚͡]ً6,d]+\ǒ-|{p:1}z-z*cZxɛfXztf]gjϨF42ׂIJҟKAZr\fgٶ3S N4qЖA+nl*KQ1K4:L<l ͈i"$-UG!kcPiT$"~ahʷd1(`Rgg*+\Lp%gw82M7at֎ɣܜk&צ?fC)]>2OX2^W 3q:~[=J t%ATyQ;YE~4f/}׶ G;Y_eivˁϨ[>?6~H#uUanB| # NA/DM>7/~ =e7DaUB`*DN?ˑ'XC/{aIrUe.k(OȲRsΐ5,D ?\;}5qU+tL;<}6>W0x̡ZZг; 3a )C (Gr8e:38M{?%K} {G/"`,'/ 7ΥJ4M){G'|c]N얮'^x' Q6~Jz_s ^mH sxNMߺbM̮}Ϫn}nzz7QzBp{5GNIp-O9KSoأMw ^_8mעjsJO17#n:7Y2 %`dXGm ZA4ScAA&sXM3.y УhayPJ2/ *Vb3OJetڡCCYK'Yvq_^A}s_^Z #|#isds7ph4&4F$ݔj$Ysq{w^W]#8}nw=ukv:k2gn]߈:r*C;;_\~$.o5bcƐ$4gӀdmSx(ʎ|EeGDQQPTQAHWC.RkGH3' 1 V˘[({A2R1I< {N !n$'QA$U%2׮sTK2Sv82g^9K(}$.izk|y' [cmf21'ZjkgM^^ F'2+DM!T I6$#HkKup>ZCrc#By&sJA/cQxB" J*2RAz~a4N3j9.u\~noO.ɺۜI#ĭu&uϻ zNo}2 TO8@ᵿ)Z^}*4n@'jlfʠ)eh;8)+>cP8X+6-)[Tj @ԨDUC*=ޓw<2hsqϳ )F̡f%Fѩ & ܳ.-R .K9f\|up*= 0KD`-<^`L }$Ur ҅`mo黫Xzcq&?Q6-zY^WH~}"޵5oj$\7U!9INdkpi\KBr߷1$%J4J<~tˍt#_+|޶d XYhz&΢1p}I/4ʆ6R\m Dz` h}C$LUh{e(R5G2d8J [Kl ZmWX`P:y^&چ.A j5[">y>2?S>cώYƎh/pIZݘ_j.>GUMI׈Hd9Q%5AdD -D=KzU{RoGQQb*&%@[&8$$Ι+2Ypk%FDىz(q`uIO?_'~sl~-mTy`d#i}v?E.Vx D*yA8,sAC,}#!N)Z$ hrnGgd*XTqxBs:Gb$xwkxU={sC^qyYU ݸn wӤT=#Fu r Ah+C62舖@I+0LlTC# #ÛRdn=m&Bۛ0dz'r?]i׍RWơ.+˲k@g2e@&id$!nSpD_8#=w@/H;2;g{&0r:ZO13&F($$c&&sN:GRl<#Z/fd:DPpȅ`I3P'Z{c *WΎ  j:6cpcvnݳTggژSϑ(k)A@f6%x EɮGP^@$>q.K8JmˍJ!}V>qk,=9`;9ⅥXJ{JLJK(՝I9]{¬@b{meL J=b=oB,.CP񌆣1NcT!1àL13*k=pDo1Xlo׺>|އS\ WR%eCT(ed#L^ :D+o]WVE_t\}~`={Yo{Jr~;qq|y9e2#gXHT ''ZV hEbecaRh̼)GY72Z-,? w=N\'eFOU)R" SNg]Zl i5hv?Jh:71"DHo&]"ey77v8~ew2ӿY1Il2͊( /Vu&Y;*T~8oa'j7!}k6:bdx9ܑדaeTYlI!v!;Rbw{l7l ~| ^ iBdPn/Si0o^ y0/r'KF@7ތn۵mض%'D/맻li gߒWvZ@bG4nǵخQfDzy?Y+{}[Fl0q67[dM+2϶<__/OkfJuon#ˡQSWҭU5R#o(.mա,bje3keEO엫ulem(xIP%}$L}_yf8_+3c|ʽV[X% kE>s2 w]"Ǹ="3ͼ1\M:zD#Ot9%crΡJ{CL.'n]$ic:-b69MIf6!]})aWt>_>RLt'Zh,)sL1;4 cHyɄ@)/dd Pd~z`x:`:83[T:.&rKA31jvqL$ڨh]gΌB =GRBL|]U/祹*>;,Z#d100OVjt7ƎYLm86\8fQ޿x>%Zs`u"&\4%X昲̅8wGNK۶hmwoud޼h[/cy{gCV/i"w; Yޢ[Y_hY_-(2V'?[o?onciary}qX.Gis,JfȆ F/EO8`Me;zB>w.no;=&~S2 {$$/]`cm30Ũ~2-]ߥ3൥o)]uУAtKw9ܗ.`cuҡŃR 0-6w M>I[gnʻh\Lm- &Yvn57}oVec4N7/w*hC42֝7B1Y>r߇e83>)Wdq-YHifVŹ嶪"0ӁkҀ"lidO %!Uq:lպ]QN/̈́h)Nyok^}=+wrk[8lu6Lړ!&-t)ᅯ窨CfHg!dIQcaKbʱ'2#8Ka59+)GЃFL.1ٜLD$P Zw#c=R IƾXh+c47>υwlsokpONb=q1R2Vrf%Vytֻ.LWzA{+2Dm:), 4 I0&j.rY8s̢Zw#vҊy,]M:EmW]=x< fBZP3)`FD-+:.L GtFfEb Vq !DCX@a&HEHgr1FDaRkRȅ wϹL*&&Q[%qh;9;R\[Ӄ9+{]e(YP(R+CIڮ龜p*<HK=l6;YSVkb1gIe2gHEHDB)Dʳ5px:K^nNH/s*1 " S (?2XD54"9(BR\ @|xŸ ]t-!'?"Dl{]R¦Bhb^eZhRܗYIq9 G\N)8fr0)gDLJ_XmuF%{F2ڣ" Hl,"%Pv+$ 0/fZb=;Ws.Y/c$߾Ȥqk _$V%(^,f&`9^XMd۹$_x3'lĦȷiT4MU8kF|4]1GT瓲X^ޥe5ڪyi`orV*畇Y\HߞV'|~:6>ĚmoDɏ( 5aPNɌC)'nHg~|yX£VEV@H.;)RMNdDɍ(TL+ͺ㟦t(3sUnhM8=MbEdiKfbFCMoXGM0]8esuvv-`9WM}}^<[4OFPt^}{Oo''NxS5gyBќtK:գ+mܪ'k?:V;<o*Ě#Z:Mb2 H>j?o,fz!c'ի)i9E5+Eb`G ,o>ifM{k܌[$Sl/> qQZ QQZPIՎ($tf%^(CL(T9(]d"h=܏s"(IcIlAB,p`-yEskK'>Aώɫ/X7-zEL1}_zwMRZl:l))8+j͒K@EY)H);~$NŗkҐ5$1z*k)CIK$謇'L@]u&rIaP-2Wt{NDޡ0R͙"qJeL< "Yv ăsQWWԤCu_2yץCw\\ԫh_Fc&P-n9ƮO 3|zu_jv~C2uvYSv,Z%+5xCBTr9{p!*9DR3nKe#2R)RD(qTHcrd1`L0MGw!N lwNN P*IQ;m4=6{N'Og[C.zY?ữJswpjEPi559kUx\ۺ)3@{+җ34:y둛[n<}h0;ڮCiqXN盻Zqw]YR?\\-/{vxAč1_l et^'{7۹uRd-ֆjgS S:FtcW˞4l{W,h`"\]Sy\{,K '5?{vKuz;"t&MNaF6M ۟Oz&8gDIs\ѠdVQ&_/wLF$Ĕ<-^.^&1}Z+b@YE+^nNrJx%uDrAeD(m|^[$zk%.pN\$Mf&>9+S"k [l*$%%InVXC 3B IR5Z.+ȥOr2DϸYJe@2IɤI%*\"$[ hqiY$ľ) i2(>TH2sDXh aF`m"2Ş)Ө!zFH91w!.4 1}`a4)AVhh.lt6k'œ gĤ\B RC@N`F%ViCZ.S&F( ҟVQ cm=JZ2 T4sIAh^&gXtr2ΑW6˖z FpתC% i2cuUR0Qp&Yqqu8~֌ACrhol֪9j+$v[>QE~1$jgh F2Qhg7f6Pubzb/ 0RB3 aEuL. maa8 -VkPVRIОWIPM@t4[hspQS<##A '|(*W,%^f`d0%,hC2VG,Zڐf^G-XXO}0"̓ QAi4/_RYN+ 0=)^PfʩF3wt)A SĬeB#(Z$6c`@! OWI%I f ,X .Z I,1'Bi2_@# _6d31gK,a=O6 vY.)tx@y|3.;m7 ݅k~0 /V/Y/!CbnMrd=l6.^Cr QUQDl@Ơ) u4ڀb(4vIv18&`h&bD`0ciƗ)I1ά $N` bX[9 i3XGfұK \V9)I9nXZ&mYRHl N_!@OM!J%rrQcW'E]%A/[. w`#/z $dyYF!Bi5 YV+ 1^ B9Vczh2&།Ynm3RTbMDr|BT1yQa8I "1Mwf<̰Mc⃍ɗ\J+]-^ mzZI|4xK!#.ፍ@Ӧ4dd :$W{HTb06PH}h +Э.H#38`Gߌ+w V'ES Y_~^jqgE3#md-ŀb$Sƪ`|GWr[65SU+pm i,{Х)G;R{IyH!%:.p /s#TϺ^H@/SAʀv04K p[O hg]+y 9.z-| b*!FC'56!FtJ߳< (tg BEȚ\\P]qcm\g 0S*fWbdUzõƂy0) !.}p$,RؒNX)h^ /tXII[$V̬$޲[PRS[ q/ӱ~i,e4HX6 ئ{ PQK0ACj)h\7֣s?U! @6d$ @"J:4~))ZЀ1iq\]4GMWKzKgmk_l2pht$kv1g5=~wSH`xV#&6 $ߓy?+u[\Ѥ|+X= ~h ,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W/)rƽ# 6Ru\1ufPpEʱW,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\Ji$RZsk\є~+ZȂߥ z\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł\W0#No^peh ~+ Ă+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xpł+\W,b Xp5ꖂKw L֫iF+z魗*aXn.a]3~!Z +٧Rg_u2'eꇽU-Y~]ﮢKi9,A+&?\ _|w hX-)=|ݧ罘/bGMgCMx>fG_n0 _)kBn= (\Ózd\.^oTQ5݈;!X%硫Ef(!(H߸ v6Z X)< r㲶$&MLq` .gr$ͽqڈ,V}bɃI `} @UVg}^>lSII tb6`!W0,fۣ{znšCdWt)h X5~^[oH U%WJ̧ӳ& JWtnZ-Rb{ʷnH5 pi*] ?ҙ6:ϗ~S? A~[;z 9?߭tF98;?:zPif=|@MdOowwf״e3nm͏{]SEn*waq?mV:R4" xKr bj!BId]^{x[_o:ʘ}es/_ytvսWӶXX$7w(]vqupG:?2b~t~ܶz+:~HgںҶ^z=v܉R}ϔC}jvNtm b @SI\6g\pPЏ/ޢQSl(#^tTGÓ;~AZlE+wX};?xyceM?G :s3XV iEa$8x9h!T4`H;J[k,ϖ)u1X?_q[[X @J)"ȸ6VIgTU]IQ6c5aL. 鬎1x(^CnwoLJt_ᖫ>J헿퇳emX~A۪emoٗ6;u 'Tk_Ǻ~7?>_fݿ삭c. Dd.eókPv/.|4%Hix7I~,Tx'*Yd#fõ'Wpq[p*#`Fv_gxzV[SAT0RUwٿ{p\? 0^Y IuuYyeq1T,K*OjX>ɖ+*ZKҊVim2ҩTRNz\`w%/"e%V'XNtŠd_\VNs*orG>BV'k#ne\v?Jkk;.uHgk9kW3/ٵ}pU甥ƀm&գP4gSINbxkTIO:j<98-cq|:nϿݎl{`S,+1K^l M*ҋI75qpݑ6z{Bb7Bxˆ}B򯣣7WJ5ɖWBZ7V۠tl#怱|.?Z~:PEm65P}`!/7eYzg`Cn! NHh#O['Me|فge~ .Ӎ,J\%Sjim| V3贵v_ XL2a2W6 ooBp VC.@); 1=<_[Or/V792|2 bZt8_ypԺ5uUmepuLeo^&߮[Z>҅*LKӵ]M 8mBp49#DPMn/zSl 6jU߾kczb./[7K)iJ_HBBVfeJld7mao_r%_\CX|>lղ/G)$NF6%1e Euݔg~[ 광]RSLOmk=5߃+U"܋t5[M77[cJN.KN%/[iPr|xrvy O>.e=eezyי^x>D6g֯Ǘ^41K-ݷEa$E^$D6&}~T/ȗń+rtvtNYI7p`YNLR\LEerޖP#&)rvGfG K/@z4Db6z܈EvX?m dkņJufu+>8s7dsW+*HŁg3?=;zFCGo-Ȕj"6 lJU&-v^'sxiݽ/ڪ0g`J0C!.RoHxW1dr5NxvWRǀ+seOl~ MZ[ tfNceq}Qc?{6 Cߏ{ ;`AW?b!x~&lJ܊bXdWWׯ!dm4<~~v+rդR2aZǘVybڶn>B<'#]b}$>d(iGx#Ӷ=ذY#yRrskͫ^mYks#{(YKAY_^:Oq<']xzJ9q?iȗ~|imwhSa/.tnB讘hS7k $Ns&9Y61Z%M%8%NsviӜ_Nx)HqkTTDMP\B je:FbPcA9,"Rg:Qu JPͬo1vLܫ/O3[{XHm SDsB䒢.Hrx*N &3WH&JJ 5e+ -?T HG n cF\$qӀ+bB^9qS<k] J-&'RO.շͦ>U<۞'($ F^ZJ-nV}x$8G4J^hNu%:%;3x&|lИaQ5a)@t x#uHXp11@I*_Z3cf,gӅ$T]h B3³GҬc 72qsvUE?*n@x,rM $roAX ?AcӑrV6ĞL(3Q \`0βJ{%@ hLYNML(J>&)sXct hbqֶ`=YF3ƍ< ج<xE&5֫]Ah\$,z0 DhR(Qׄx#(]V9[PȊbNKljqIoЈFx :q8ҖL2V ʰ&:VI!rJ Fb7@l"C-Xzx#P*y4#F<ɝ3F,#N5^lV,&%EZX/A/zq;ŭ *%sFjl+zg딬hN1 zzRa18UϠ$Eu^y*_qcF;7D?P~\ͪPtptRO,aUAka(3Cj{}Pr/Iu%r\%&M JRg"ASC*wR-6ݝͿ]M7 fZOuT [T!UUk|ɭÿ_&'Iyyǎ=' k3m-P[FY7Q=F##ν[`RIeQ}&U '${d?~]`: <EY2qDmVdy9'I% !*QE .05`8$FBQd."%eKu1$%褭M$nQ3)PkhǴ҄ҥ؋sO/>ݏ~{I5u½KIB^qugt>4>eLLMv;mv,ֳZ B %( 8+j͢sѦl)J;WХF"tA( &Ky#qh:J+;`I+{mLSXN oy:h38zh;գZ>N/W0#i.(UR28hY= nC5irU,oiѥMS.O۵ +c ̫1_ %Qk!`)n 9NX@א&4;o"z6(OއdDDS+G\PPܵ&* I Br `km:|-{EwΓd.ł7Β-Q)hqJ&ꌦ@d 9> bkm㠹[L뺅‚gZED aZxP'r1S-D5p!De^Zè"{zCdP4/cğMys+g./%T\42P E\}܍ /"C奥[#4Q6٘)!bR Q =vXHe) Vv 'gǦޡ=IiDM֩;)89-2QF'X1)Lo鏪ڳ )W h~ig4i-~W5qfa<I_'?= kސLY妋'ZNuwR93O{*7 xUof17ʋM$lqG?}x3W N ٴzGҾZ@~RE-܌ơs0U7̾qQMos<8? e̜z[2q7| x,B."bΜwh IUpF ]Ĝr&~UUzym(eY,V-eisf4]w>ʩg:'E|OIvKVUwP:rAxz:i9"liVanj7s.Y"0h鄷fZ858`5I?(O@!;Ntʍub'+̓ F 4IAĸM)~ݯ<|$K;dɠBxb#Gu9[՟W7_]#AɃV`k<  >8=W۫eo<.aH'yLZHj6988!d̜PCBk(HE4[!)j]vxˍ@6BJSϙ5`lq`[E%3j!x`DB:QP #;+Ai`cHﲙWP=x[["\nH/wIkUar/kl'~4|6^k>]w9i]9D:o%#kjk<}ijFt3#I:z0>Itd܈Dɒ M,iA1E-,VĞdzۤޕ;N#H.SZ"",`ozF>h-Ӗq: HF,;yr8$,_m6fyq\N|qc=ڹlG5wOמ}3e!"'W`DALD]:(aqҐ4b?ķmA7mκ+棳}sIJQw1Yϛ㍵ gK'^7/$|&f$ ew+L/;[ a6T_\i5Kg_<9ߋD7rd d ~s+췌oǧ7|2&|ID*Z{_نO.Agn-#.yuO^]:V^_:A+ܧ@gB53畠)/ 7v_pc]kb:_տT.Z;}է38'7Ggo~<'.N^J~oVh*Yi6 6jYY*eRmaS&d{2&nR]r{_x|qhcowGx{#_0x8CY%6vCD+d19ȯ-~k~wFᱭ[nsH@l1B gKOݵ0c[ Pv^$aPTk+2O3pW87w<5J3/ }oWaMίX{P"|Z+[,*IY-RI[iGik?X\=҈[֮ZФ\휷8:)jQf!l%ɉѪ8-%h&(AVmQKrAW-JHNb eZXpX2.TT.Z5 D1?ڏli wOw(RR(OvJkDNk1^i@7J.N@b5crIJ[׌R8* LTGTI B@N"G-b!5Hf)e%lFZ+ $eʢ~n]ºlmh`@jY_mpQȐeqhB"n9mjIiD*b?U@f)W8Yet,yrh.5YJzxx/dp" nu1#Am<}S\oYxc*QFV:uH )SmHq\ֆqsٳ*K'jjdU9ΉR,h:xѬȵ STBf&<4}dSa4n 2VM`r9¯P#Z)dpH"Yl]0S;w|VIb9kOeM*\ J{+[lڕ@ƨ%gE2juL:RȦPB@*т7W*KIn L%Dt Ѹ jސ XrIѣF٭(Pd8gԃshR3])X7n0`)b&[T<@cјenX"BՄ61ce :24 _Gg*){u+{ңu% PYjXJKI.ȡ2" + ջZ(Z0*aUcgh؆ynVor3E ] * $(2R6M(pe dX{R(U(N`  , _tk=vVQLFڃ@z=-5lDIBmU{CkNMACXF@4rJc8Nc;ia;‡ r ` ʮ˜L\Ρ8tR 3Cu]7,2~J< pY3TLg2BQ.(ʨqƢA(8tfI8w@8M>)+]7|4-=͋@%AYa}ڃ'76T n*|ܤLt+Qџgj@"pa[M }Ÿ́ H7Vh?]D>;:o1K*!-VXD(;Kb f| R;ЋI4 QB/r!PFt3 } ((BLAS{U$\h D;ܱ ,P< %HAP!Mv%K}&56x,nZZ"/xwzDMȒ,B\{ڍj:L$EFi9DPt  XtDd*jQcy(?) >&@JY"JȡhlͥV0]1V;,vNCHZI@:;͠R)ԤފOr jov 6L[ieT`LMiX?hk|e7}hQ߯^ iώ' smL*` aiFh%Ь58)Qa|w9Yx6ZRќGբGl dL Ly~10-'Ez7$%xKTlЦĐ 9 ^ndh-87h4)Qr3*XAZ`АH**mz x""0fA==t}V 6BfuV"b%*cIB\(\!Aɟ&wQ^V2^SL R-1Ԕ*Q H 3^ցNi,+~(]:IBcS;SvVMƙZicMZ B&T?3NFR !Wt ZPY[E):9>VKiAVA[cTBmE1.A  +@(\4JVyÀeg@h2\H O8%0hO BI{OY(t }J*S\IR[2xx4D7֙ 4fc)k`Uj4.M}ʨI8׌Б$`ЅKO]0q%Gd LBָvGtE#BQR I0P5KBOz-nZ.7Yg j PH(F B yp˗(UՊы aѐf꼔S8G?IENE;(!),?%@`%;J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%w!)`! @?y%3>K%ݚ@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JV $2(`P@Z[JS Do@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JW D@7q8gus8%%@ V}J b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V}>Jև \Zf'}in._/ ;#zVgyyqrxyۧ œz1s2K]rqR~+ ,gqvR>qK_̯2Q|}9g__lyq1-OǻBb_ n-pZ+iه|5;{s%[/Is[&@ًy(ȜҢEYghWrAOo/uWD,Jo'F_/o4d^){b9 oQˠo}!Lm 6}]DN"P̪?F v/wٝ6+[{.4O >-fM=klj)j_Eym-:5_}>u%LxZ| M9R ZV 2~H&7zȕEO `Fk{3rZRԹ }ju?_pk\c>ݭv(^I  Qs*E W[ ~=~l+1vt#w\=ao3\n>,5Ձ "AHam2F=ƨ{-%:QwoS〫bd`P -C-ط_z|yr1į3;ǥ?" X\=]FH3kvmjwbuY}"YnoPڂME . e)xɳ QuhJ"F*B+QF84.mdF qȘ$YR{/kޣVMӢCg?J:4'E 6 ,d/m]j6 +a0^X-Smk@spvMKi8nùs)9(ԟ޷JUθel\ьu~y ~rN TyR CY-K"'w7i뿽: o:m#T]ߋW$8JSJxڔ>TFIV`DedJ19Vh"o ibv^QX?0L?5-\ ?髺`Sƍ!Xض{Mu;d74GQGc'^ Z]Vf>T{81MU|.>.P?G?!{mm;b5ܕ4ݳ-bj29_ZҬAv[SNam;gуaYo$_ƠI%;p 6nш N׶\?GTvM[sȜrݜi!unU-" "]^\V|mBd6[88&=M+#<{> 3%뎺{Fҵև{fف>7ۜMp;޷hu <0V~8`8V{~3l)oc6'gYE&#iσ}`œefσ(ߜ<\_xV5!^~OFBvv3:'O;6.s}xEh$~ESd;}vY~C~}~C 8|=׫󖽢ʼLպd ofQGy<\=1Bj"%_/ uC.ᡝ%@HAAe }5DN bM^D{zNR /nol-mh6w'pFOűz 댟"J9p~R/LRTd*Mɾ6f NS9J+D2RIq/o3pK PK@\Xѐb.5^ITTDQPF\ඨs֨dboCp[*D MO:QJro6/if8~.L=%uK:Ti'cD>F'&3\[:*)5UUp *jV饚Yjvvyfгhw>R C.W#SAS QE:y>[84fX>W9,k(@HHzbe4$&%t8()YJ#cFrJ1,,&b)³{,} w2ln֧ ngfa>uAFh8M` Wxt{V6rO&h9L=e'!=pg6A\OV3 hLYNML(JbЦ0ba1QKAbq*j¨m{-ؽ'EFьq#$-6G ;°\+b9\MbFR -C*r-u5b D4Xd&(4C8$Q;W 0֤~ebNKljIYDlGЉͼ *)m 0>h2,*NkaUh9%>Fb7@nR`)@BфL$wKVFb9"Mĩ%uSqEb$-.4{T)Q6RÝEfd9iN湶tL=./Iǩx !; a@z<"_cFQ;;{?>2Vsy?/nԘ?N(%=~Nݪ4Jt"ȼr?EWx8r|5?E˓j+Elzу#ʏ?%%e:N <2bV:.H.o3PBw#  ;dLfAS/DaEaT.)LYfBԤjB=ׯPwK~^WƒF=;kٲ;h-f~7i ~Ani4H8l[P rԒda Jh)pY]5dҨdXٺKCΙ(B#$:$W(M4q4dJtSjize$e$גSa"" P[Urp UPB^YΡ,HK,$Jÿj8hKR>1PFѪQIkt<+BphE֐^F)z7+UvU` IN0,o)tY[NGg{| E8bR`ޏO=|UNJ%"j>t** TPaNF˄Tڈ<:ټ2|0RC=;)9^%DH5h ƒFBR X/db1!T O>*6 Qqk_nn/7qUA >F 0xubm  695x00/wKْco0}=9^ɭ(wryǤ/!&$QE ڃgz)9qL$F h2Y*sHIn]:ik*K 5Ƅļ&wpqP07Y(E"2JH+o(. OX8}K{N7fE h*Qx ֆz,?Q&%hǴ|;eGwΓd.b$%p͖)h? )LM@d 9T} b|%Z'eG$aVi-6LBp"炕␨F!*.h‹S7qx>f# 6:/. DLEEyP=+1'ыnyiᖹf 1 hRRlLĔ1)([\QWtJ]IycSО4+mhd>Z7l1>%ɐg3CġD)miQUTu v\+ʌ2BvFDI(u{X ۽,@;u{ס#ԇޯlM ir9ho?1}g,&6r5A~*W"C VYͪ#g,w6PE?n٬m/q}%'dwXۜr,~c_fv54ƀs hӜ UpaJ0a%蜌0fx_Φ38Q8Zlqz1c b ̵?61dVH.[O#bmEjy3m%բo<..Ql0dQCKMoNL+ )΃D6ܨw:E\̓7[R}`/VhenY]1PGo y&裄XUÚrq|vV8 ąJȀ$D:;;P@e*0P(9N4>lI6Q%Ҙ肐xF-J&V`"Q cp0u# `T",lfpR9urΆٷ 't^%mgF1޿*.uhӑ+{ w{g0 SxZ#h]?Fur;:]DwWZ"hjݿm{eݝ7yvzF[?~q>4?CE) ;6-w ҜOt߻(Gw9 2i)}>$m+,1h~ё@]hd"0*mF8=22$K+ g"Ƶ\~3 gWW氳x;X YW( /mY(0 jԻf6qb ^}0͍άh%N*#TiFZ8Γ I^v;EHxBcp%*:uAI+a܊'N%>Qr% PR,pn ̙CNX`"3F+B}r`V{UPړ"IO1!߳lnf~L^X5ö gBZ(rs' ]x$,uKk9Em0 f^!hK=Q%ıO h9 6`*j4TI)4ǚIi\[;`3gP+ r=qʒ\-V麗N;9D<,ñ.3s:ukDG7cz@zE8$=CRObb20pUYα^gi7x kZ36UbEOin$}[`Zh\9 F??d8^&oM]Ns9^Ffx2![뿾^C0LpjsQݘˇjA)()ymA T#@5w1!yAJA]n6Ѕ\r=Gx"Su!жm[ J&P‌yPnAkx"}D<`bL.1qHs'h}ψ!.wy0kd+^smIUV &[ۮ.sPG|wdA͋R3X矾UߖT;7`q`LZ\%-%h2T5hXx%h_yo}bGh0}ׇwvfVbHau_nb]oGM V?|q(?76Cq(?t$0%WAI%WUry\^%WUryJ*J.˫*J.˫*J.˫*5K-^o(7مFɤI8q2, #\EXFy<}^pz#wD}źӷݳGVl5a欎B:ǭIQ8GYk4VKC AT`!CW$f ilyKonVXYv@4SԼiL/Vhd(/tN)K '^i ڨ%2Z`Et#xFU(U$6$Mc\46h%!hbxwEDdn bA3K0#RdXKͭ`(3 )ƘbЄt戨 )O0d󄘾bֻТg] v Akͮv ;K]M]4Oqxn=f{cHE s {δn׊̓&JPcbl;ۧ!7{ Xj`9 ›5_W/z.=Wr?g*؊ٳ!48'3L?_Y+dF\ ilБ;`dow27Gw?_ev!O/nEs H"0T=&@9cj{͑)"1_jbzצqn XqXwJ=~ʗOF/uy&?B %gPWv`T=֣U!tI+oIm-7w xkoЭZ3h6 Z /x\9US㶎$4N^]9uh[F=6G?'s۷i[p_zow?j]7nU讱nS7Hk(Mrym Z%M%q?F9[4iz-!:hHǭ'Z'!WRQe>6Q%9r JeCe5P21C7D/(18-(CNH]rgWaVa[w>}PMBoC3JHtrIQ%UIǨG+QD$Y%?l ^ _Ehe˜DL(|PvA/a)̳?ŵJf}RI_ԣK-m1- O:w"O.ZO҉{lLݬX{=o%T>}Bs#ӨAp(}y>-sKIYN1&/*}(z `zDp"B, VҒ9%c9RLce), OIX\gdٽzwjn7~7vx,~X `QX(r3q&W:(ᕈو`rblb‚*H$hSXb#gIP1"qDZRvYF3ƍ"65[|G&56]AHlj,o$!#2A3A!pD/8< Asa1rV_EYds)8R"rRV"r2HA"x:qwA%-3 7XITRApZ D{xb7@"AG-XrG@RQ4i*tbP#V{*~Ob\r\" wq ,^4wEUDuH w5# ^JA:%3Si-A.B.Ec!+,q"lORTޑ{o 06 !-֞*q7BtRO,alU_HymmUCiɺ׊{|IRl"5l?Y{-<[2vͦS\UKq{ 7:a"op  yMwCD2mp&%gs}d.@;0ʨ1UwZ#q^1}M$@-B([#vt]符#;)?YMGѲ],g8ڢaZBLFD Jc<~ㅙ3'vlZڢ-=zX;(w;G㸬Gzr ji`(?-΃if./_&zq:_EbS̉xp?AT3jP.(9DS\ߞ5 Q`l{Fp"S K64&7ZGrb0iPPޯk4ukY-yfkr;}h߯~d]7Z%ۅ,Ajp;he;QͽOlI -1. ^rG`]}^ #p\y v)T@? ’<,h"N~$'dzO:(HB)ŕ#wȤc\+o9%ZiNa4uuOڞ}JĹCC4n Eַ2e!צB-KHF"G`J-bJUqbOs!Tvrq?[P" Fxg_`֯y3[?%ur"|AX] Em)ӖJk;1v|Oc mկ,1XllBjD1J]m@|j2>o^w3y!9$Z2cgoqrъ,GPo]✒βj 2lh/Ž]|j lJTna;%t-\E;"wL$z굿';ߐVt]w v?֧g9h'bb fV:~WcTv\NgeZ]ZzGTȚ5W_ =4d{Cۨ7 "J@)-6'(9qvv'{SZ QQZjy'4 "؏t裞؉ҡLY2Ąllt}P/%'(Ib1Q2Y*0QB$tGk33ځ\6k/W ٺ+Q|  B|J~]n6jb37KT,qVԚs @!H)f߻U6 \R.!t"*y󠍐)Wd,1PГ|erJ|Du<}|[;7Փy+7'H91z*i)C=h$Ǜ0U$ڂ `>QR'"qrkH$Ȉ0K}h$$TOģtR@ %'(‡9u-wpiA{!-ޏMjּMkKU;azˠ!uVbf1<%j(5RE\ҵDՎES+RSt QӧRQI}$w**鑊Jz:Zzmָl\F;e9eQe AT(,!)&/vVK! #s*>R.ԅ2t; 'NTHdtgRP ˿tuyPAaݱ<1ͻ m1L, d"@̝ebJxX^/FXԟU&O4IijgBjh lrGr|zuչpKX_eE% |6N/\== WJt k#)iI]>ca{0l}߉Юsi.ʕq1Uͧl-,ZpQ_^+V6ܽ"LKzXvͫj׳z ~fq=4Dɴ?.]wլ:]Y(ڼ|痊_~w+m"5YU[S./Y%鰳%ve@ p5PjNLk0/e˧=n&*iwX6@j)` s,NhFaZӮS&Ewgv`k|!aG|!ᛳgȟ1ц|!aG|!aG|!aG*9{~c} bHZHko)^ /F"ݮN[w$}'6w-M 4-`BLFD JVO 9!3 JIzx3&s2B׃:Pu6Ɂ܋aՌؘQpKoGPuf׃q6kfM&_,|kwsO9u!~5e緓Yc x!5̉|وVtqA jVE%v1%7srÅ+xGO8Esm(@6Mz!ѻHRMܾ\$j{O{Y2@kJ?u)N82URQ=?.JWW(f>*3YRO3Io 5kpaoPͽ'lI Zn(sS,vyq|!~9fB.aX KiED6~ЄX"qr\bJ)dcF&ZyK,pҜzhqt~ZHS*D(KhhrƍHVX9IHTc7vbJ-~Jcg4Ɂ9>JbEz{П8Mgsi-'obFW&wQOK'U3_xYAU ^>{KT1ގryWf 4*J2$ƕה+ş; ,f_X|V~z3s?qÌ@ F][R'|:!xcm ;fwujѲ6_IkKYP M%9AЮtfZAkJ1(gpu2+*?p} ܰOfՅj*EyL/?Gd#"C/.0㞦[ikܥGtspߴK*GO=ݟr涸!?9)2%n>MKr\m%S/EցS0#mȝK+sz_L]VkvOglCRn]jytQ==>qgҧ1{B‡~Ӭ&đ,j=|]'o 00Lv̷; aCha K^uM9e* 9qD/ђ;4='AIF:d۪e"7-a}EX"oַ['^YEOwLo?Lf_' S>n0O92ՑڐVºn^Lh}z-z]EҚ V* 4roA@DNJ8a U[k4OЊ_,]cr͓%lBM !#$0rǕ(_LI{xb0X}Y>)wxesk[j-@v0 ó+ m֏xa1\g8ͨxWolYf^].IÚTD5Fs={'GYiY ȩ[ey|Ks!1xڪ$ g 1^$7Z:FK)`8$BU͙dryb@5BD!v&׹?:ikOgÇfϷ]ma9ڷ=E.Q9 *eJ]#F K0LjaR :Z{)[?-f;G1IQ(Gs E [zpԃGiDD Ak%KMH1:H\:GnkCԋOy#k໷4%|;,p. kf}I!3lgNAocf7Ғ~u5~g Ůi9[m߼~c 57ў8zG]9y>ptIed!xm}vگ/g8^SʼnMϠ{lJ.dTpQ(<'N((J%`(:gJjSħW9AesT^hw#кybHa0˷_[~- \PߋHrvL!h4IQhEk!>%I$B册{ w AykC";^8Bp -rCcad:,4@2!C؃%v ߹l_CH>Pfxrk)Ve_}Ens1 ~ǘ4^km8OH/A-S H P# 8d/x<z*zz$u'6"q,D9426|!D1`D{Ӊp݅Q ^=,9~EP2 Ft#Di5\!&ĕG5GF#s}g\4L¬Тg] -Ekn/p?Ev♁Aϧ MS.pS2xK8AZ#º`4#Pmbb3+)8 U׮_xȕ<υ`#W4jh*?qV8a׃/9R`<)\a%D}.M*Qx_7yڿU~N̎ʫq|!(J\iJR݀OsLadFFTCq)"D:`?2E6q(WD |rkwL,ן\Ofy`u=3@=Ћ! CpucU tیV<،fzsCj5&n2=/^LEC-E+ky>N*~6"Ɉ 8mmJ%Gm[b%ӳj,W1}8Gw*nd.H [uV [o. *f |!Faшዏ즖mH{EͣȻ𹘤4`$7&z6]EB>w!a߷F:\yd@rP]7VP܏kZiy0%/Q@Wu\qtA;x5O+i: m&1p63f8Yd4 =jv1ŋi[cˊх}kYMsf.w<hߓQ D\F@Y,O;6.=(/?ȍtbN/+nӏ7ԑo@𜾲pQq "Z5׿GDxBB7I|rb-2eU|4{qS!WUUM9j| hjAږ@u%K$JQJzҦIL+\9,JE=xW'*~avJM~j61L$SqSvhR V;ڎ\pXtf0 잇VzM5 WmSMphSu?n|9(,+a@ۣѺ%gKhd-&JimOh{Ϯ1Nŭ_%2/~^}7<~P8cΕOX"vg>2V(1g( ]*Pr nqb-"ZR|,)qI^x8?zf=_Eh3I/a֏XWso,a)Ӭt^FxIAʐ-rؚ2OGk2{Bo{"I̫Q>SԌ=I"M9 "IB%޷/֊ok!)mO:Z Řݨ|q!T6FKG6G)&Xf].VXkS=l<^ŕĵ׺%'e_:#,6o߿t=7Eu}{mrQE4n4ZVIM zVC];dZ>@^^x9% EQ.W$EbEH, (L1?xqUܼIɆ]%lzۛstL~Y|-^?K͟$O8]\?ޞ8]}܊ڿ)Nni6fHԎ2zcRzT&r1(oj}t?k2bw2o=L|yi7)z~燩QwSkڼXg_]5n]z[4a^_ضu.E\,H'Ftu3fe:+mH:ِ*&3 }XՃz_5#8a?j\NIBP^͘'M2Qr?=j`ѫ0F0on q;XcՅkJcTbEbqdl6 ^_-zj`_o~ʵJ$eԖ$e*u*Cxy 69G32_+Gw>"iXs8yG嬽VfN~'ߨ[)pSKkm^Rb$r*m/H!{ ^ }{4`1d| #tlj%83#r`~%Ǩ/e`*;=գyzΦ V~ډ />>hϿ>|CE}ۯe-izZk78+eZWvJIrm/977퉿zz/L;xo%"sMS7FQZfgWH[MvZl4㽿(L뿹M7@]t4u*(YwZ|przG $ڧh@{ 62׬stfznKZ*߽R~.zC'm?b )r|wJһ$'WgͽNw'_xFJϵLzd6?wqd,үōvucGU= h6, @ʻ'f>`Zfyᰨa:Y)qS'uQ藹&CEI[%9]L>3۵k,Cu/ɿ~ZdYć[G-k/O'(_^sJ__ G_,OJmc.%oZҢ S-MAeNΚWJdrcHys]},9 r>V 1:*oBʹ)j. $~knuٮbNݓ.Gi-cMAHkGlZL-"[zh /?֜9 TLs&kg3UM*՘-pOKPC2Í#sYLk%O INIP=„{"Z#<|W YR:0vp-ЗW%cM :IF* B3{L>,FV9Jk141{' 7*#F9G#Lg`ıb:Ioջ$U 16,F(@I3jgxPH{7gM=DUs#oJT[pNTE) 娚:'a}wFs5nb_YkNҩZJreߠ'̵1|&*Y˾+n#i,I{mE"6*!E-C2)_UHllfCa b搃pʬ 0tf͠~=U B.%놺@JڨP/3 jv3)GK!(<4݊ F+ uqDgzOOLg_T"@cl4q, jʲc!Sd_܊%I7Ůn*~Eg,[Dd=Ѡf9XP<: \,.DYtT|]W¡{'Q4(B?Lw^%q,cqd2[ɢ6+U.Z)*QKl3mr嗕2)RGZë64(]ƫ&6$t=2hIU:!a#KX€_ T:!@\m_ Zu$PHIoGI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BIo 2H C\B~ϝ)B[$ g! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! @,T$)C!@\_  q$H)%@" $!@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H}$0_ S/q y1$I r+$!,V! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $C= v/X#5{?x˸njp}PwSq0Ϟ,`l8*q ڍI3/p86k,n0Nw0\_߽Zy q=旪6&2 |C><_Zd>'haҧɸ먟J/4q>a W)|4J}%zijT!ͪ"cOj@_3:rBMy2nC;EZRTswdE:Z%*r%ʬ#`ldʖKl>Zf6jIЪUXY-Ea`F v-[VͤWYUcB'`>e1u(-Ј^c|@+Vo@HR_rc5_ϝaѧ:}n-śa\zh%Gfmi睸0 uZwn.>vuy?= zLFwf耮;\4דxVF-Pُ,"7Ŧ흷Mek]-xx%FoWY~R\^K LZm5 6Dhck|V0fתAs]`uSD,K #7z~x 3ަoneYVÕeG\gj;e3vv\W^wn8ޥeͧ屫Sͮ\ܺX u 0+St%l a<`RUhc$_{tDKCng"ؔ4^ogUGYަ*nZdCeЗ;Yb< -7Kݿ^,;F=3YJ:G*e<,BK"?sϟy30'@uK՗DOYP%u5M wm% . e)]V9zK;IF-Y[]qq=Zǽ:nzԻP>798Sy>>FC?^v`{_^s1՞koEdl~AYd__V%p+T Y'IY')%XD qSAD'/w*gP AykK-L{lN iIIKV.(K-MQ8XiLBG`lY%iK>\ɇR%vWӭmVU<x:HKZ'm8ϔH/3OA-S DM*FX^T<(tvQlU=y6Cq,S"4)T_,d$3`"Fx>)j^4O1]~^"5wtH %K(#Io 61B# q͑⒗8{˅u[ieja8chP2~P|Gٓԇ/+ QՓ1;r5)/%KER'%FZ$ V , &DTWh/N&f:;v5e>T%v+_exRy]2I["W4jb,ZY4ʁPeI% eUX"Meqv75ྛ /TÛcx]5cu=學~{(Jm4R}aTFIV`DedΕ1%IdN`"o[ҽE}ƅ\-3ʭ3?0L?0W1́)WnAVMM<-U&i>MnhKC7'cև-&& =k⧺*Rdۋmc߇tW"[b)rճ-w)V/> ԝ٤7k ^RaVΚ$_iƠM>pkZF4„k!PK&4@\57;ݐ<#{x*.|&9)餚Jg DB>wu[l cHg2/9 X]͕^MYuiS摋ߛETĿ2aj0N,B瀮sHWJ=u9ݛ;txi^wW"q+{/ oWc1'붛8#yǃ}G~[V"D~s*<23'%Xh;{49!T`WhG=Id9s͆rpEhp7. \e9>!N6zO=K~xʼLZ*f,QB=/GeEh=T) к )RVPJ CUV ޽or*RtݤzS@³UW^%ZfqڡV:ǚ& /LϠED6:P@tjomK54M~dMןKztsx*Vo+Ymmr[SSAe=8oThYGQɃ9N*EL9:1/m  Kڗ{5a9=)'(cbTO W,MrB@IQ̪зfM?[7][8VuA]xHE X[9ۛnx5W)h"(V' @ jT^S pKs6t:i$J")!gSG O9ƾ4hcٍdVWS"Zzf!mxM|&?Y5vn%Z 畠ő99??~kH+yLBH 0Pp@m sMMvbAS9]}f_GΩR̜""A[ iD&$E8W,pks>U4Y\{5'@P^JqkHH%!ߪ$ܞ.n:9h\FݍG㑟nP9ht{.(nx5!~-P_PV|NG~ԋ7of`?vRGx/ej} vRdeZgYqP'_7C*ʳ Zw;Eى?ׯ#c(Ya?clu͸/8 d鱓ؙ`+ JI=)7SH~zB˦Gn Q%k[7:֮UUsfjD7խſ>em{i5ܺZ7XKq~*~;mGw7n -h0rݵ)d~/%oupTޑ .z1L4ß68|s}I6zruvUZxi=mi\CW\\.a6 eٳUZA ]pP'%' !$L@4筍 ],'tb33ڂ\E4}(UcBWo|eb^ { - h˦/H5,DcAEUd @c@)kk-tk-ʓ:!OB^ؤg>ɉbyXҊjǽ6B&\%N=Ji6=^}p羿Y[ ޻cM%18*i)C;hÝ Cc "2(i6j_L$0 (ə8\]1xΒVݍsdށ2gXr! bZbUNk:]Y̫v}%U#8HY[\!J$bԹ !+QPPr漰HoI,%|H1(OއdDDSǣ6imS\;*i- y !h7G4[M[!50-Kww<8hmSP,xc$TDn2,M4kWb9%h&j@* 1;tb|3cge۞ <*2%. B aE$=`F!*.hv['ѹc4W+̖]f#!yw:o.xfR⚊@#s*;+vo N,W11f0z-%1ySBI / ǣXArE =#h2BH*s%x EtCQf5qџ֜\|L+_cZQJ"|gۜ''<9\p \%DU"Szԧ[5mV_+w|]*c箟ݚr'CSE\;HYg<>~wo&׃Q x<!uOD|\DgpqN ZVE%v>$Whsr}9OpG==(k" TQ FI6q}w%#byU2Ku]9y`J-E-E-WE>L[m3鍳XnTR]Z(R|Uٯ|ݩS't@3R<ٓK{ +=tM~k(c|u/ޖhkĥ l[%2$h" 84!.dyw+ULB)ŕ!Ȥe\+Bӭ()Z'N/es MѸ1Ed+4Ν%l@D4HMTF\8 hc2qj(\.:?>=mO_p`'fvwѰR"xWI]>(M@+SSRSDb-ױɉgt &'m19>28[(E䌠BdBBRHJ+odv@Fgyj"De|j@ QHFbP<5 3uμJ>.g/\!xfTQjlbdg6쳉DZ/i*?+U&B=SBC"G۸T6"<B-[]~IfzNH kz$m-"C'"&  (b^xĭ:DfqO{L:FXV ue P.vzuLyo|!Ngwj񛦾_7C.?[5m&_ 8L RǛm^iiǯeA!+OR^MRhK[H) pAFF#o!hKPYعhIP(Z3=9: 5LsUA W/+Les^;{8Vͼ]l%I 6D ɾxp^ͱEGcpzq Jݨ>*n0!BBqnU<)" 'I`7%+czfL:uN<v9NqZvPLʪ^G S4ԱuF K@B2Q*O ih9E,2oSjT=PCE6raM魱\.L)9!HPܢR%Jz7֝g?u0MykoVR*)mֲlҲd{Z\Pfݤ۫%Мe~pϳ=[ 5/w~K(Ŧl㡡 3GƯ>K5Á Mt!(&a/6.!+ ƥq>*{\G:8O?|+eJ3>,y: 4ˊ|dݴq&;YΦڣ0+#0aAl8i'IgM'4f0.P^xf t=QȨ yzI# {*Z9.UhSN"N$5ƴV̱')R|(+4o7'mάݼ6ɘ#<ٰ9ss4Z R-MO]P*&)8م!"Z^8l(a C Zץ>ܡ4EM֠AkM.$D'3>eXj"qMk4&j! 4d'0 (%3E.:k%Q)b%+ƺ;&*п_B8¥vP詵>[y K\Wr @BYҁ\IXB2n;Haz4eNz_l#] ^h .Y:kmk߼}kո}ϩ>' =&K`?ġ>c%οwмi %u2(0ljU.yPI$(X J4m1M锠yTG#1p2m=ux-ᯛ_nk yYLd/g/+":N'|h`QJKI9[DZu Zʝ]"dYŸ]څO;kT )*Hp'+Fԁ7ZGZ# ]֯ ޸L~H_V_ޘl6?nCjQ?e\ov1޵6O]ap !!8`]c@U3߯j5Zi34YUU.E#ۤ@:)3g:,1Ygp4@F8 GSG0g<ߚ!@#! B,OЏ~Dw3#p<SD-5OSzHW/^/*8}g 20b8l*#v;= Ƙ`d I4!YjR_MЯb?j&PT۲@d%TC.] ?fjj|]6̦uZܬJY`˅!57(r1rse/rc7M-25%C09 HmqFu3P3%[]ig0m8"%!:4(kFkn,ll}Z-Zjιug2D~X)c. ޻;o_GsHZ4͟ON2_)U]7W3c=;o~)5S>iVr՚%\wpkgmw!ZUDM40O4_ [E`Y>X^bݔ;: FK{>_3U>bz^>=靺u7z.aJ$538Y?CIXB i~Z|16&7뫛zV.Ц&iqOJ@ yS>c=;T8&*/G WwIƴwɋL4b:R&a0nh0y7Næݖ99ܓ.`cǐB0$餍i ن4ujLFc`7f,P[)z7hi,7x3ud<1 5|od_ĔE>gN; -vR&i < րBx }vRG0Q},U(]@58ɑۋɩt0Up;Pڛƹ߇e91k0C6<H7EPQb2TvUs*d I@ uvg2~' 'IHpuIa.UjlE@I;#XrfXXԽg}ժd8{|OX'IprʭlL-Y{?<\U+o?2)+y&-9pAP) )c@`FsVg3@.9dY@KUUfFvXTӅ8c_]*B7ƒG<@ 2^\\Oq4ֿŀ6\]|to\c3a*OG gE.]4I d{ e,R UE:bL((jSFLyɜByT䄙%*kjlf OLcոc_+km?jnO@mIZ$z{Vo8A0>q 1N*q6"fN- a5ʩF>$\xGyއ'jQj^ Z>{yf랰`̗]=ӆ3M^5k1x,ſl ෤\DK'2%Cd\&ѧ$컃%#wd6фy'F:7zvBKdV0e4>k[Dzod( -dڂĘzSp@_)qM{w@/;ޓVg5c#wLn2ʦb)IPbOM`r"60FA9*Ed&&ȅ` dt>\ ʤF(E4[W/vc (}n10/̚U3z?Itk@dIEOğQ|ً'b !B{z:zFPnKlay^g]A7emq>AkYrMlӃqzDn$mײPކbl)#qN_bܼfn+v٭IڷwtM̏ktZdf~loyu{@mr>ewm T?4-嗏M%~N69_9˂+zImm^% ַ'l)AY8zFoq$O@u6YT1g *k<*Q8, Y s:{kiͅ:$ZGYqU6J$X{\"FK5\CO5rni#Ow`™Pm@][|.58z$^[#|'i:|ѭE;*Lʃ56SP#߂ @j@ӖZYb;*u;?"*̓ ل$@)*0S"[m:sLQriLy!1ǚH~8_W)ߧOs}YT|eK1QddG2% G(Q6E2&yPY꺺J?}*ߵu:}ڹn[F}lWŤ0H JuF#x-ӈ&rJ Ǻ.zV^|u|AeN SrM'1A: , ш DL.6a&# ԂNXl&U*5uD$HH]$**< *23ihr3SW ./sDes^$5T"L,>.gΩ5uedðXتp^tŔ “#iB`Q9+@iB!ePcÄ;f3g@쉕fn_IP ϛtō3vl:m7'ϣۤsyjZ㴳M0~SϽjjq }KOZ)eiMB3)`ۅP-¦1xx1I*tNTr?M\0>jzQkށ>V-3ŌJznr ⼹f\༹B538Cm!\݆8k1 w˛ x,J1VY2ur=#ԨȲ*.&8㒉[vrA62xuiG>+זߝaiN>yM:\ML= iXOE$q:W&qlYiak~DcS8;1󫮰A/ |9K,%_ w d)+ d^,,o=+zfqpn;:z37$p5?nScVK6=%yMD6Єh> Kjh0+c4VH@>AHkgJtc%wٻ6%WzJbc;&ٗ0*I.[=4M6mZDiT_*2@/;])Z'c )"L"6( q0l}=hA1J,B"ƞ mrpfqVaq=;Zĩ#1Ǒk5t~t8-JXa?mW(:-ޮJ3y5g qKnӢ$JђdKk,)QZez'*`cڽA',3;k$B蹋 +1: NEQ= `58,,$Dr=ܠrnS R#Zi vl[JT0W\m^гksW>DIP-iZ?5))"q@YM%HxU%R.:d ˆN. NrKNWǔ{1(HX0"*FEGC@U׀-U"ba@CJj13F @ͅl8^o?)޼&֗_7 ?xɑAo0 W«bL|Pre#8J/1X,Qd ҂+`)$|@x3Gk-]#;B+XdLHI=R&TaED ٍ2cHQA5>E2*Cz Y=wh[,c ղ{/qz(jdr~a?zVO6(/?į{P9+#SGCG&8tQ9__&c\̑85CTbrH8 h@I?!f(fT⪸ʫomKb9}菎C}yÁ9 IUp;o"=1qV{tHbaJ{!^]->u)>u=Oǫ0$tB EqU#ෛ>o'8+w?R ;VmYC=z/{ŏC~R%$;ouCfz_zōr-*@v~z_ t(I㮀S۬P+2`(-e70ԢTqa$17$hr|bNQQD2aX5KJGtXD&tN!7&=G棛î77CzΗjcf GnkwTzj60HԈh`R!sOFva) ęrӘG@tq*j 4wFm6*GTLwd$ n3K9^uJ|oZt..}5ԶV:~oW ײ,5ey`Nx;cLg+uh&6m$)\h9`1Xq)\E8ǹ{yg`V (z Jウt0MB"eJmlz#Z'ŚOtIX]FGiy?}s/Zqh~OdZ<Mqdkr4VWϋKvw(6*gFJ=ų38v֖]%AV:AfĀ9"s(ۭ.g(eԪq=]DF $&rAtZnH4CJ3\jpZ; ƕQS =,qu-\Yj:RjKteaw1yъiv]{xH4<`io>%㸾E&a/ҳAK^x~r1(kV]e;xszHKƅK5^6j y ^0,!g`0MӾRR9.YG`pcPg%/fz z=2X!73Xr`:Z .t0rN_̓7?O]j.#8+֎#G-VX+&eTS/<5z08r 3|(O?V9g0 NgI7 ?jϚunؙL6nZU}߸~UcsS]%~NiCU":yf{bC} |>|-vbT[rPW.3*1b6hܕTb+V[1^hJ8Rd{D[N׌ihc=Fb%5C1bTktJnj?#vtk=Zu~TGvfןƯMlr1P 2eZλ(($sQ)e*e#Lc~Gi/x ` Z2cN(S!*s(05K): ,Rg2DOLE80'Ifb΂wDðq3Ug#g7A/7O-eIKzSwWU9z.,'-w"x[2Mx&70|QYA+ﭔRQ1GTj" -GGL+?^3H&y1>R C\6#4 өy (쨵%$Oɓq/7"ftGp<ʒ7t( 3ӦeH Ōt `ՊrB `8 \NWYoC =Dgܖ>L 4LDfK9xb.EYTG.B}Jt)Xc)XvKL4LXQooȕ9#:rE, =]ՄQDh,*¤ڛR*tL֕V^Lݾzs?F.S]}jh95c0FL)'T9Jc,Ukb~ndH;m T ` Za [[L矚/*}װQ@+3)J\{l=j9L| H-g=Sw![/^D]f>^L`c?վ먩1I'AmDR}<|L=}K"w#M=Z@`xhKU߇j[µn,7{zPdYC ^zhJ8fщgpG5fhSGw^|jlf/ɷM}7BTzA6.ZYcoGEעX?V?P¼>m6yֶ(@ct;uӿ>ͪ h {xrrǢs9;Ǩ{t1x1uEo?l:_\pp Ϧ#/־|כ=)zLx(< ҰW~0ߚo enbΌM 2U5#+|?{CU,KpYSʷk^7Y5[kpl ^OZߑKh?d}Zœu~venCgf2{5?>4u s׹,__4.e٧m-ut6[/z oSiS^UGx{[(jH{f>o9H 5ok"mxddTѤ j[=9T ݘ1&HԱC`g5n"";{pog=ݮxvw_o̸e7~PJIKMmK6%l"Đ_II4`Ƶ ݤXj~H<Si ƥ,XZ})ĜPACf JN-OSl/Zp] fVGȘ ,Lĺ x LG<ؼ7#dJD!e!x0k5f,`k1hFs+$̀79;*O@.gX# AJ瑌& lmombY_a˷>ɲIo=tmJ>Ry1IYt:"xFCNbatla̐ -0B}3B)rK%bdFYU.dFÒ1V& }e, IX\'dl^֧]Eէʵ|AiZhJ)ayOY4$cx$6N0+D{S"#a:0- 92Kl<,tLSvǾR[gzzhM H*IUimTR7LNtH p9o&*Eg ,d@h k|DQNp p'B.OamwQI0sD(Dl$ {oŁfHkBjE4. >ޮI>8;_8>KFn s&FhJi&< FՓ7ق)'f Rg4e$D",M}~#%7I q8.#$HZζS FQL+bTIICQFK~Ju(x8@[KrJQZZnKBYa.%9B|pRx AzgR&T1q:4o6^9ʀXV(Rt8 XuFɺ==O}%,(&PJ2d3SD1Jkq`GƒQqSZ8g,^G)K1Gcu@fƲPHDڠ>۶s[#[a2}NX.IX ͥZLڄKlA::bE> N:U2#r BX-ϗF|Bsى@s>sl&)bd I rnb!5\gV_4IUfپ((^|"-Fq,׿.=`i6:6lL;cN)ubSb Kn{L槜RZj)ed6r n:F=ǧAKdPVoTr]qR$!STBtqdy};Q'N']*KV>ϿKӃ5SBE.2f|PLDKQKl<R2"svIT! UV_ߵ2k笡5Kg 377=7P 15F+7+6::IFl@$"e]|K˯..?\P]~su! ]x"DQ͍l4"#*:-یd3FيI?^(kDk|$/OClwQˠQsi45hx<)/x&x8+d^Y l \Tퟣ@Go[G^p9;in]cJFUCD ,Ҋ6-e5gR"RiO@ %H]ȺÏ}(FmE" 9PD ec,:TZ&%KI`s1FڴKzܐeJoѢƸ5:-3ͩP9dѬ_Qi!#CHdt_6yNK%>~G na*D\걾i&q};&^(1 ͓NYj6{[5w\B,YZAT%Z^)SƳ>Yqk>'$N*<2238&9f:Wj+i=/xGk0^< Td:Upxq.p)ҬG %Yf-dsg`ER7q"9Ѧ#7.i9s< e~n= HA9;UϵyO~sS"eIHe2ԄĖxL9M 6~a- Sx$-#@u9in-~-OpC6$X9:P'dQMטRQd&eb<d&2kyە[#i%!"dq1ͣpt f aAp =:lwY]VȮ>Ou d:jZ]?N'IXP> .ӫ׏㻉a)2$$l-@}v /kth MmoX _E<5" >.{m3TTA[J_RyX/8 ^L!Y9+2XHWW dAFAEE珣9BY6qft#\^TtT8 lkgu nFtkp? -;)#J&encnlB{u ܐ 6ndɆKuu=!r'in'cW̚SY?[a\,fG]&NL'$w?jBIp$szf~%iv+#wd41a/k>Q!x:qHi9qdChٝ>=]O~vBW0L6ehT,R¿"9×m\**RId!Q8!>lRmmj/`QRg)) E.s˼R:n%C$̒RdVdMCg@N(_:xٺ\|tL0CJ>Z5 Ȼɬ[ݐOB)}4qZ(s, Nǂ(b![Ayfbq#Ժ7a-=%:~R VǪOtUQA)OxUihiiy8z6d(ڷe"f-KW&8 <k(hX۟ȣWï]}Sa]Ly:}*C (u7m-T2#ƲHI: LIwh-h˼Sn޼DRS+d3JZ|* !MJui[Jj6+tҶ{5p2sfrp٬'} KثY{\kjHyEüޫOLe:::F18:cb:8d~w!# R]x9!r$PI +$s}Aԡ'>;%}OKZ6Bth]2iEKC1L̔"Z i%9vv$.b]m I( zDʖEJ25(?8as?ɯmaPoJ%;Y"knldrpT"7>5ӣmVmI}LlJj0#.DPG}^>—uN~Ƶ$m{\teatƫIUWĐXo1WS`D*"(Ҏ&[oh:g}Yٿ/cvlnjhf.ٞSxXkM[1DolX:3-/VѺ~x?\W/st+OW{w{t=/iz]*&RM9|En k/+7sU_h^{bۃb߱-ic>,{sl'GaiFeUْhУ-mP 4{M{l_zzÏl4_tqv>{5_OFӌI)3h(;_|u?Λ֒/f/~YW^h<]ӧ/ƈFfêj],!.; .:Nn8 MW;Rgխ`6|xy _d'[dx84[AzgY Z|@'( pQ,@x]?Gs.\P/Fl|XL|1^oOfcEn'w/On/7Ihe)|튺uM1rCOd R0Z*Jۍc̩Hŕ^*U+.77C[%J6-}a|K|m|_ [Շ\~=[uD / ،-$ÅЯ7ǮE<-RkWni8N68|:D<chT_?[f̚v:֠?U}XV4:"_F}8>OocJ[/X(c v0JkiҌ[c/k+8%-^'x݂ɳH"ʕ*JX`#;LL6j.F#f*'3Y2Av9h/[*VA_SQN2 ![s\chSc;,>V kNk,H(`>3bkEݓ$c&y1vH1v&tYȯ=b;|G h|rM_ia*GGgV/n[7nCWYǼ'mH~g)JQ(Ak2֜ i-$M]VB@uєjVY49Hp%&-br C-I0IPURijeѩs{oYn8>[,6qcO6﭂nJxEg,C԰fD]7$ 4djra9,=FdS20Mjt/wBJA( ' ,&a JDj:puTNh ڒaR8KI8[uP!p6-Ɣw} Vmv>k719w$7C9y>`lY̠<*=5U!lK*&0/U9ejY-@i |* N"F*KQoC0rܤt.9jy}6V ,-gC>Qӡlc촍**6{Lh>.Zz|Yq wGc\M-뮽b?L5ށe;̨!bR9{HV.(rEg\+$`F1¸+sYZ*W!}z 5-5f-Cmjb(k3d͹=3NU&˅]sa8 /[~678vӒV&o hsrrnUJdbHM\I=i߄Kz+ y.Yi0`5rK6&:Odg"6*ȇ3d͹=cŜ'f8q֎=kw{S95_`,#-z΅#[9b6!H)oQ"Nz|XE,h%p8fYD.>M&ɃujjYs{FȣΒ*is^k^"ʱkNzh޳TfK6ƕ,,_k$G1>zǮPOwTC kWD̓aiy^ލc䭳Ob)AeUǩ~4h[٬CVs~G g̀Cq=ko2 oc=Ny5f_mٯ¾j0}yݴDy~1%UFt q(#h1`){e0\X%9ΑR)LĻށݙMSG0^n>;v`".V y*T 슲) ft 5T5F ]%  Z N\&A[!Z+F!)$ih tY=Tß9wd 5Zm]}zW\%%WK{ƫKi0(*]֬9#A㭑Ljbb([86}JV.RP1 ;Pʹ*@E+!%Ϲ$N֜o OsrHT`B @`G3=B>:*AV O=X,C[!it~89uFިj5jɸIW7ΓǪ!JJ4d {Yk'{å~}mzH'J$)SdsFբ8fj-dm({N+ d۲׌ ![TN){3!j`)hk{;hVC%v>󝚺3uτ:~>WU<}gOLOm9֚Q*EQ%+;㪫-X,ǔb e'.&vstHPyGu|RKNx_z}Oe9Ǘ"tXJL՗U싵 T GI /dgd֗ޒCO *)r6)ب8kTMD.&t>6DIS fy/J~Qi4L(3$D'߿~k3':hJlFLr hcJ) Bsl V>En<*wSnk7qrbȉI뗐rAo3OJrU!@ }:%m FdǘlfroO9GLn[{ 㓟+5* 7歁*~ZmYlqKuxM<WkiqML󪺛(RML`ꮮYe?i[s cZ}1 UKV8=iڝśv}~Uo4> \:!`y̎1jU ^|\)&Dž_ဟ=sN*5Gfsț ŭ0Yf~t |=~L=Й~.^܌_Xyynun6qrJ\0Lٜif< -(f4iKnem/fC ばOq; W4ϯfEgf0z1\2󺎋^eCOǺG*}sL p |^_*ߩ(qAYޗ E3R=OnO:G|cMp=1Q?csV$NN]7?qK@]N˿cc x Z:{nj5uޡ`ϛmvFU瑱uᎧT,k:ؠz^Sg!ocLT]"/}s}@ 7tk3Ita8Xu̇:݌gnsZx` cʃ\! )Є\[_X H@ `>/omn7QGWj|gMrIH[J%Y${Yq PL  Kr&),llY0t>< P.{ZW\Pr [CgVp<ɂLYqrX:3H)"?.^y#)C19Ъpk8д;^{(bhaB¤5qX[axf6qmGVE<.{Ĭ2+;KT0`nQ(}YQͫ|mjX۷E}k ?/_4Upŧ&@e5 WU?w /=910 .fdrHxGN A 1nRn!p*.uM.F}uHƐw2}\wsҧh#^:Q  ߎ6ܸ>+~$U,xvQHr6l.Ll 1Thw]lnSCjxd\xtZjjHdcXfneSDGSwlpa pe hN[lB\:$UX>]N<] k͹0h~[s^<#rn7I|\:S&]. %2fhQ,Y[%mhv@ ҈[m2T BdB*$'yCBiOBR9&ZoYw}j/)"L"6( HJa}=p,pvC`[t}{z<vVe'$a_Y;3(RdKJo+P-!?)D?)?W)4*"/$HSSR#= 58.,$Dn󞃼 QmGs ֑aY lX+`JrD4vt GThv%ݹ&q6t}Sߢt[ŞL |&sѧЮ?zBY$%2(aڂ*AFr~䧘s^Vx 2d tKV)5A֑@"\2#QHY !x0k5f,`ZFL#ЖhFs+%c tp#Y4Z ^2qiaxj юcu %hxD1[}$:Ex=2:%ekqW'ʵ *T硯OĜ'8ɡ]e|xKQ"x\:%ڒ_Q8gm6^ݘf|+]ah*fAI _h{e=.d 1IN87::c1bo\ǓK"]/$U;^;<4ۑFB`}C|Qq|/=MI@vr<)]y#Z/^ޡ*7E_&u]q YG3U>^y- ^ :\|N(] ˻ C[6$Z=ͮ֬O^ zŤݝ?MKZun'onJvo绽k{o է/%@ Dz$& ǽ8?" x,(8Zg䔅9愈w-u5AT 7TSSʷΨr];])LwE`8p^\ DG79?LAfӺ+eP:+W .fޖF\pRJ4 ׃ޯ% iQ ݸUJ'z 42'M.B\PքR`bZ# J' yj/nQqx~8\=r`YQ}[Kd?rqoÔFq*~ћ$IswEBP,8-/O'"J5QQ| lx^_ ? Zl܍vlw n-nQqZܚtB5ݩN(_JsE0*h)-:=[&~0X%A Hם@4kj)yiq[]!+8kHZ-ޜlR\7RK61G6Aj;+rW sܵMٳu%g5g_Lzʬk If9S1ܤX⣓5zo˦i=`L&}uaKPK)!"݄!azu;vw[rixF|'TW3nj.`{'3.ќ;շe-,CkԜwuݾ!wAb)w$C-ܯ~N mL')W Y-)3'Im2НVwJzo~JT)WV;M6+uR2ݸ|f߻ 0 M'sMY3&Y m^\8+5;y<#Stfbjy۴ƥlj]ߖoǝoكoP&fvYmY+8Wdj]yS-g.r2"m-GjzaTgZ%߭t 2Q̤|A %fl/ZO1c.: z)f 37Ō7u9f,3 FUF$Y\%% O l mDGjmWI;>}p~rT9MM'?ϗ1\|)\ 룇L'r7~/.]j~_xΙw2۶ff-.,E}(lpv{ZStaO~G-%7` &{\[^?iz|u: {myܦsw_i??z%_H-KOO Xxa(y}zd G^>>+ b_`1{dlya"nA+ UCN(|8凜&czT2\2li=;Z<8"-i&kOv#;N]rŧcs%ϤmR d(p`ZWoU5Y5Й.#rBDatLO|T 8?27-mڄY3#z FVh]>. *h:|dn1p9:f<56j s؞Բd*;ˣW=x{cnZؖxa;yUGsoӟϼ5_'W^sD8Eid.\}rRU X( ǞX,f";P eJ\n&G j+ ܤp_͏i7c;]f;4kT$pJTT0+,qʒ(bs\)-WJMXʵG)2Q@X F$ػ8+\ڀ2؀^Ȍ5D5ef%u-*)`0xSqBWxfbko,Nk_/^k]Nl΅ΘZ,m9[̛!o|i UcC^ikfH.]3be|!Zf\aLf`)ddhiG =Ab_ h gvٙ/_NY4mĹ ?|[MYg,4x2Z.2BrCJ:.4[pJo@PpVWF1DK{),řlMoVw5T퐱J)G̙@: \[!$ti,QJ ՖPE[4 Tjn ɢ&d"zsk@9f@57ԩ6ς4-[A LcgG51< RH Q0\w(scPK,Xղ_ Vd@:k@S+DQQ+D7ueH^6u$( XeWǨb«qSQ@ !)N-ގ0**E՞jVtMջJV4hZ eLc `5!*@brۣB6٤#j| VcF=*,˭u;kW/US Q,03ƂE.‰]Fe_}3̛9̰ ךWy?- ^aC#D="?z;h˻ FLfHǂ0RP8rP >`!ߕ,fUH"QcV}qLI.&ec}rSAoJ#HTDMk !Zqaec]5~OyI|@֤ /< /Tl t`uq*@FusJg*ΫLDvF:984x&?ɍ$gqP'YTIQ[0"1w@:9h/pC*).p k GQOc,=J pK9Fc@HAk^oB f\%\a ֛} u؈`C8Fzi-w}SIYd0uol e1+0ۿ_h (eWg;g%emQFhXVƴ`ʫX>#}0#brӉျ!(A/Q{n9Ѝ-$:کA*G{%T*d@u!uPI dc=Xܪds6'7lE]1("Auɹ a9w(SoH˪+Pô12vdr<nŀ\:Xd ,)clchpVp `N}76k [[Fj҃cMz@kFyHD= dbpU"R  QރJ [, ?x{$rMj8d*Xk*:leP/FC, LčrX"Kd={$SaϖV*Y:vil PHh; T[S(]xD LBS :; Xxk)DdC J{ P V/qz^N̪ͪW2}UQV1y ~ o[_[\hU7McR nv[ٴ~kg n |hq_Borߦ[{ ?-o7jjv <[ȱ';XdN/H$ /G$+})"l]$VD~@dVcwC&ώX]N6Jj?\UU ]y^]ts/+gGs/3n ^{|jaFqn4>;Y*WgA NrR)~3N7͜Gq^gg3:QB*(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!(!=/I gry1JH֨gVfQBL)LWB*"m]~۠$5zjV!??vw'c.Rnku5L7+gU[݊R2@[i."=9Pl@-SiW&򘶒Vi6Dw78׫/AI,8/>yXiF-D{֗ǗڸtXkX_T}z5T:}Ҷm&2}st#2x4yO.yfsqͶ>rtt:oo&;K'޸:e->cN*(&[XX)]Gy^(\չiO O%>!q Gݑ鋿v[]jk}OݶTJ;[]vavavavavavavavavavavavavavavavavavavavavavavavav^R; la6<vX==Y"rՇ?N]J*Ů*h`#ճoһ5Ƿ?>iy/㺽s#`9t-F2jSjЫ5P|i*`Qa<}&Ч tDd~ϯ41 o>+>3;J6ǐ\~{دl9r/o5dKf\ITK^i$誕rq{Sw6SG.`lVoC1I'bHK"(% 9JJJJJJJJJJJJJJJJJJJJJJJJ<;H/IPAA XC%~DA  J~جnUᶦq26A@NʗVx**dJOBD@֨<;zRK>;fYhn9?x}~Dhxuy:RͪNq`Hˠ=t2*oUo˹KɼKI-~*G0O#fr[/:MMp'z9qh{>A;F%>[o,{oI.|kN$|4:Ey>UucoWrM2“dKb3%uju%|V7+\R:o{wmdϻFn2dl&iRJG-gJa^]8sxDǺ޵ ֩+`Pd}F)e 4wި-HO zZ,cp`0Q_f3u}@FNK38ϿoOMqG}6:需qxSqfBų>sU/O'zS2ﷻN5j僳WzXoK'Y3=*&%vK]No oTqz듋#nSۜ_\Qѻ~/ۑse_w?{WƱ X >0f K6sO0E2J^U3C/Cdž-j3S]]U]}Er~n#IuMgSX-Zޒ3F8d~瘾3b~qy1Myy'x+]GlN'q9aA o{뵛l_|a^aZM-x+b[t~x>< ||/OKçV+,DB\ILy~x<^ܼNZMjn̕Ϻ7#\p|jXgp2o<o< 8p3us7f̍&X 0 >]3m E|Mw4)O|SM ̾qȇ WfVyk;Y6~ܹ{x.)9TE{%||ȧY6eL~Or7󟦮a<r> mNլ|84-~HmqH# o˃:Zt5}wˈZ7uޡ٨6ǻӹLVۆGR J?q0V{F3hmvTr^eJEԣY0j| ϒߨZT9!B;MZźn]6. QD8"4"q3٫&&hK/@4ɲ3ߚ.7V5lΜd4X cԌ%J8%jCHbt ˠ"+NJ`ȯ*n&6L6ɞaiOj'"0ϲCuU+UnE9;Y]X HcQBʧ񤞶#VjNyO[V=meO[V=meO[V=meO[V=meO[V=meO[V=meO[V=meO[V=meO[V=meO[V=meO[V=meO[V=m+;+4,yUӒ0u OO}USPu@Xf] yt/1 4۪Wu_G#QZª)K$m.rh j1 .R;0l8{7mU٧vg~P N!]5J481IL# / bR6URj$&jNΆ{XUNNSr2' Й ,>`&z^co;䬮1kQvѬ,梽OgSJiM:=:Ht&wR) fp?91/#Nc%q7[5cg٭U:ӅIơt M ]ݠeW9Ý.7w[Wa66?AXwA5566TckǼ?fxZdB,眉naYc@gEF3T\"jf#Sc36 2m:؝ gQ\q1OEkw&jm۱ֶ-حPn%Fьq#$O,6Ok#8pF&56(E Y0f%wg1x#,C*0' &&0Xb&(0AGSb11I;ױ>l8aa-Q1 _WqF\EB#׈9`eD3J[s@<“ KJiZ`$'NOXשFCpvkĻڃ^LJՋcH{mhNzU^6RjYmhN)zzTagq>dCv>< BӤjkh' L7i5Ip=GJִdML{dbvvO=G-+yt.i6.]GHO kQmUZ/h)𥏔"Xg'kLFɥ=n0*nFfjErTx9;_nwH?Ouk9Gn=8i75my L+K湠yT}zqGh\(GuFJ&"# <2V:.#AEo3 9^I>qq>q :~ f0 J{vI ׁW/^eRc-i Qfx=씋Q{AO(젱l5zg0TbBBFD40x e *I୏$2L|]wpdjz a%|#C4ʹEX,Ԟ|͉xZn`YtW.Gb0Β -S9dD+q s:"R)֐]%gu <| u&NjyJɺsU@׸x+l`. %T412gR*6gыG~C-M{M(5M 9DŽq4 w6`R1)6Ym\Q 'RDO$$S68ā 4a$^͸NE>v91deLίO-aR\/ټ" z.?ױW0`Ta+=/4e5x= 2l'ZN'1,.ȇP ޔiOTk9VW}^Q,ޤzKyv=i"|X`M$lq,Wߡ=\鍙3o"1}#&:ꅷH,`=oVˣt`C4+w.i@w_6]xFsDo.Gfv5"fݕ9X@/0E%v6!gp*]Y5, mٞ4$'8YO] $y@@I+{AxվW8%ӨNk.ĺZ𝟣L\C{1mU<..IlXXnRu7f4<XnfHj ؋YMinQ#n`=/L!(keӆbMir*e}YP4GEr94s1M[=)bI,R\i<sk-+ʑL֛&Ɵ/e={y^# {ʄ̹OfָAee&["(U67هR; Viv39dq`3~h@84?=w?n9ӭ=ndvU!<xb8WG-V^ G|g%Wxa5Y{clȅk{{b}lG?e>npi"* c.A]ӭ~C6a]w+7k c(TMh5CI2c0E=?BY֔SRoGqۡgEQ}ힶn40nf.;Jb rG1T&~wr]>wGI%ErCBeuw`-AnZ5[GˉCf{m4!Ƕ%[MU;:]ّ>۟f!,mn='=/٣畖a:oo~v{L϶%Mˑ9zKB4?oofnM͇ =tJe?+:G>R \#l2QUaCHmҊ`eږT]8LN}FAOHh e YJX=lt}P/%'( VĘi2[*od4t@CІ p`gAr,m^ɾ`^X7/OH&2BṀasums9Q*Q+P̿ e1gEYR%R3 (YjGu?]?j48KP\lٓ :2k.H:18!0b2br-X{nIKlH*9/En[%;]?p90b̒c1l5~-]gn>K,LUc~0󢺺բ2_4b>?jSwn9 vp 8cLi? T >BϞb՗9|r) sb%b Cio8)ĎIZ,_0tDm"FҌ3&$: ekB5?{6d ,RGˋ`{g2YHr`}IIeփ)$#Y:U} w4R1?mE߸$8ĥ"82&#/I C ;XL*rbIUB?-BiNg>ǜf*O脇hU6M"Pe >{U1Nszh3+d`]UϵGEcQ`X:k1\F(+^}Sv[R$A[ 5݄ Mt%7O&oKO#2٤JD]h"Abб֔oj2[SJZbM^ ǸK$M\GVI$ *}r(`FL ekDg!3QZ"3ZZ@I#1HN4&qpG.JqԞ:ӏ#Đ/MsJ\y>OC}Qq{m ^Usg]l{e=F*QC@*R*že,uB)eP!{QؤoOYe™E̤(O#b`ip(\z<0[ǝD\E{ZoiL՘꒗65@c* (׊ ! 8k!a9 3F Uu_WiwC?y n`  h^}Jo/A]i%DȅXI䜻rT6T7mp@G~cqu¥LV~l΋k;W7MbMo^Q4"ur)\ІH' D YF뮘T\yUO7P.QÊ<[3_LQ7Opɲ[px[&4ɷG3^nC~+il5MߡEZnޞhFg!BZ\R 5,WB6RVDKEmKa [ޡCЖzrܲ#Ы g\+'9j& "I\r< t,s^X-V"|O%YgZ,Œ`k:xvqJ/TIءBGGCG&&A6C?,ph<S^O  YZFP <1Δe~`Q^UNa:tT3M:L0:z.|BIk` ͛6x`OD5z^Aj3쓞4#!>KoURq|y|[_׶{!'lyr Crz<aEPO&uo~Zχ2sVWWmA}m=U:`#8͒!̠Ti9E,2oSjYPơCFC4R9XV[\.itG ArF*̢MYm}GiDN8wՕPs%#qEY6Y-#Km+v3r< G}Yhal%+%Ԍ5+hƙb蕦i3uP}8w1ܩ3?{MI(,2X_ɧ5~3$KIJ\6n*VxW%:w>;Ujh.f٢epR9>Em76kWi |݌"Js}ޖ?{h'Fei4Kcd{o%{R9MJ 2\Pc˛iurG3amO$_VqKlf"6< uXG4PW/v@E ~Q mbξc|;랷sL5pMķU@}I|F;ۉFFi,oXƘxKG޾Ѝ J3) T3-"J6k0LN9E(\o< 2FZjǹ2Db1ƕ.(ΐ ْ0!vhSi[JsO\ N*S-crOCy0ue«\#$J)TVTDqOK,!`2AI]t:H;"Cx[ޞ|,!3IQDF48RQ#ya;2GtW1%/9J=}vFVxQK= 4@~ڄ녂.%]0rm6XT6P0W/\ʤO/af#-~q?|>o^Q4"ur)\ІH' D YF뮘{(rg,UwSf1b0> qXL/o#XA^Q5j߮Mupt;܀i?ր;MOp#_j8Q &ܢ-aZoG[g7:iR\v4k(vURƗB\LShKYb,K"J- "*lyA[K bt`@DpGE-IB$H@4OBmI 漰8[DXT^/!srVpr͂륛M^~nRcIdt/s3 a&/4is4p=[u Uv5ӤSH(A +Hм7-@h)QkoV<.L"a{9HvO/ 6ϗη%w8{my !#bH`ٖɮ^^ Kqإe񔑥N<^H$t8J=n/ٶ^LgSŬ kN)DXYo5Aq&C2<1j !M[w$Ŋy&fkȞaΓ2e/N|sG?dMħCT:`#8͒!̠Ti9E,2oSjPơCFC4R9XV[\.itG Ag~ȑ]`:!Y_ep a^f LQ$dg73}mɱR*~bwӭRwGRƬmYd0;^.u[osO־JKwx"bb,uVjMn/T{Yjuq@Wx`Wdb(T{j/[ϮR;*FTD!RغZUë>)!Eu0 MPBAөegLȚ2UךU"b4ޢc@ ֺx^3+B(!h> A1U) W!A)rZW",y:H7q>@^i,-3jAc(bWG'# E+2,^y-v +Z3NA|C5YM@Y 1 &W|*jo/G 肐&ͱ'd_G]HمPb.!GdK`g7azz&n߫a)!ۅr^E-$tBDkgӶqBmS0 > f(L' 0zvEumw2j,i3Yn7ֵς';{icq ֙Q9FlCUp mJ@ƠImNJ"P`ye*hB>*pDV^C)^ P)Bg;&Ξv6oFE]~HQDz\Iˣ^[9>dZ6T#*EOL {.=pl, h0:'༜J1ZB3ۦNO$ESRxofFN6G9SɜrE$Ú:v*ɈU$Z|II 8IO +f ròRCkk(ЪV\ "Fdm"""ղyЫS1n*rƷI< $/*@0Ba:|{nmdhqTj:[ ^Ji2jrBd %ahHMB”lH&Va"loD$.#Ԕji3y@!dK4eyt֥2ŹlO\/2M>j2e+\Gr\{x4Cyy\R-B]$248Jʐ-6+xݹ&I{#f*fa0Q /`G dRUdm0 u"9XM2e+|w\ hn>:7اtۯ0ie~~d@' XE7rjo Pբ8ajۭ)'bk k[rn/kMoa:y$~>[Y}cH&TQ6%5("H%N|m@O7qcj|Aj˱T| RAM\%=qDRLrYCM M\py~]>^\n=)q{3sp,K/w2=þMݭT} ^!̾Xudʨl {4dB(t*K'`\B.gL'΁J%r1uc^B-S K m9Op?lƵ푾['OoX^ zBτ< ytJ\A"O2T8'$6,?X_rRM9S.J{=!OizY/N p|dxD6E\9Tl1(&. ͱ%XH՟`2X[lv_М]B}Sԛٝv?CNRo^t!jAgۡU鬃A|9‚WkR2CIZil5O00~HqtfBc?g6b 1G[gksXÆUggW/ںeY'iϿl>,y\Gޝ_^˥lp)m,U//VCb\bjs͏ߝW;JW>?껳ÿAfAA}W}.~>˽ Pye~lr!x7iY!V?O7͋?k\kGpY P4_ne9G'vg0v ܷj61?Е7K/sz/ksn+&޵Js#_[}uMTj =>^_wO]\Qg?Z"1ifx 99RTN밨 lInS7>z̏4 2&Or?-͏"\>GŏG|8/Y{t^3\a,Sܘo,šN}@5Mze|wnbyns#Dfkf Zڭ޲b3ޱV/h,64%>[w`n=STШ풚kNCjA֯7֕ʋ_FEc" s,G+B*8EgWMeuz!-nm1E+ɈVzDX1'du>(HS4NTz]@i Iۼ[6_xXuM6o=-Uq'񤩊qWQ ylհZAӡZ3BMvayhUg O6rj0VN p*CnyqdZ LCbѾ".Um2 RPG[rZ_J*j 9X gԐRQ#m諆:՜"hkS:;ٳA>>C7ϗ6~4kA):=ɖ"|L;P1chLU95TQVu}[T> sY;dB~'.ɐDb$ހPl[q P<˦xKqK˲OI'5 K=e:ۏ7㛬slFڨxJsWTcPbR9kd|$]\a4b2k$Ό)>CԪ ク5glM p'S(Pk̆SP+&_X\,eU*GҌC0t0a3inr .g%y}r|{.g7zf?fg*FVe S )2U@Ym|Kz+ 4<e64(ـi Γ쐽,Tb5 FWTE>tFnF켘-y*M;EړŜQ>xc EϹ0x`b+g0 1DrNyTDqq[ZEj,Fa pȉ(%XATzz&n<I&ˢɼy""jB .979G%6!"f.P):|.U.Ež]OcJ9EQMm89S@!N{(HIH^#LL7Eu\\mO)MKE.>IemVbB0X2dl_k$L1:1pTv3(k#wFk'xF1&;7e?>R6Z㱲7f>SysCڣnY#i2F7='6{>V㗌Pߔ>j$Gw+ l0{J(rr92-ūוޑnndnp pZ\тίWK*[KEFK{[ߔ-ӕl:7+TbnU;jKJ)xY+PcPjd+r`X@ho-ryYU1 ,BWL`["b%!YGI8*(2o 1VxlcGAX)ٷ%Gd#Y+*ɮ60jp0WQ~N ץF&o5J*ɡSli%(q2tHn]rZkmȲ/ Tcf'aZ" $EQlQڶH ݷ[U ]RGXbho![3+ |wE]fۨ7]Dg~Ri.e2v[bzdxPmԨc(g\2q|_sJ.qy<{<_ 3?Z +A)Gc%BM̺[\-Zߣr9nqehG ј\+Wlצzޥ19Loo,TJn w.ǺyO#As_uCK*ghp혣 fKkڽp6ƇG)HGw@!߆%7u'}?84a$I (Tejp&z3m6Vr4GY#/q*e,4{|TA*YkJ):H*QgL'Z,nL|<|~MP}+8[y۳Oin =[U?_,oG 1Wgz9luPɛpNT`qC#V=ĢO~w r̍iڃyЃfܲ zrj֩ i8~Җ, XC*1Nsalx `%u*nFd x9Qc\nCWLh =mҝ _O0~,Z`{Avla }>QZ+{0G-KUd)mvG'P',.&}cYx,#M6P4FyX,[NPiF[4",Ou؃ %j)^*jJL,C]x_b$Eﵐ'lISV)+꟏Qz+Nk~S͏-ԩ:bCVvr>;*@riK [O./^B./ΔG \۪VN):ٰC=^ݐARnGlft^^4nYF{?H:5e>v;O`7?UiO/ ' zav1Y '7&Hj!l%x%E+ğA`zPN8:蘦Cv1V].nH& }xO[tW%{Z]=o=\k!m>E Z76>L{uF3]=ۛΆ۷{o!N_yʻa6qջO.OUy.o-0lVXqVUWweG1=|S\o;[eWņ:ǴUk_ oy<0⧋ ڰemC֎.IYڽ 7,b,7gNg :FgaRFVl Pc_| :FL\TYϜ Ŀ?R%Gο~t|ʓ Մ$Ra1l6E(uJטTĝY!j~p}fL[O6zd+XoeۉtT M1dVk!5̆’GD~&զȂ@VBiM*gŭҤ-e^99Ї*P!" ת$E lă߈רkz~Lֱ)\\LNFs$iC'79wyIyˬk\:eaZ@`.M"a+d[EߟugGˊ2.>{ w]S痳0qw>LZ>=&Fy^yWKB2zJKBHXxq)d责G_/y^h Vc w{,tyMm0u[^<{k(=HYB^y޽~Y[=n~|pVzl2_4Lo,tJjfZUzqmND_:`V{T;ʷY.p#yK+)^ \[9ӦMK}K3 yS"11..1w]_h,l?X#=¶ƿԴ%R(W_g S qOvRx~`gzp_I/'߬!6?d-*wӵ WPHZÿt;%## #;|p1=7n]{s^\'f6}d[DM%KAJ]- 2ͳDd( OFW7sm %7Z~nT=%z}[vz \quex˻*QOy^M-VE:iT4QT\{;Q1D],h9$% ȅ!,R2%sU-&CzER:k!H|9*:Z8evM;E:TȴrHէKjrت̪,.ktU(UaP-Ir.! c0V! 6cVGȘD9)J836[C=@=͡+%uJi2 -}E ܆V6ͤ$2@'NU < ߧ{Dƹ Veȵ:[6(kf J%pqo܈̫y)&3g]H\$--j6Y->m\'S$~R]S +͜-+*#rHЬ c$1ڬJX5Nz`<9 ¾^k&uEo'?P x qu~BeNb!NaaOp9h\L2%68ЇJLa6x!5TiOrCFkĒ+,tJqƫ@tN(= ^m..9ax.MV^ 2^S)QP/@t[XksHSR3#LAhU<6dJ<ɠ-]%OU8ՆmU&IOXb Yе PWVџ1T1~Y 0XC1e"( 0#slù -4 ;T%U2@j) )zg ө,,PWM +S )<$Є{1W* 9bv@i2UI"I\:H'E@(a;&Հ\I1#~hڇXAS@so.Đ)v 1 Ü/p J%[ LREٔ<80)(ЙVD/ƭ "2I:@8"+%po]MEwF%BA1WXUbCK@;rBj-2Zz]ZvJ"F 025"IX^,.zFO(@H*  b.*)_,K`N'x؟.0Н Γ):J{a:2SV28U. -ebt4m"&AyA3 _XmAԠ‚BG1 $bDL W$X胕EJP*1POm( +F ЭL "2`G[Iw VgESQ_~( Ӭ(al2JRPfDbSxong"޲;8I뾵u1qWC""( \{#@ xtm^K x1( 7Db6y.e}t|QV]k)1 haBAhv)0F1C-xWC3K:+mA1#kFV(-5* ΃p:F0y pd*(.m63Aq%JI>2A~2w1`DhJa<9ye>F/z !dgZ~M;!gѝ0M`QIR ͫ+U95MeV"U8 hߠtf&= pH/4$(2c41sɚB~a-AsZnu2i>wQ;MK' 5L0,ԵEW@7]?{ƍ OٌUz>[nRqgRaIdk6H$H3roR[ }uSf#WVld1p Cۿ7 ZQ yZAr(i` 1* ʋns]҂{ 3l =ȴ/^!6C>!.wnXXUP\B^%x9* ˝{Tu`R`ZBB*@y[`jHVzdXFk lY0p"PQgmУ%%sr-+1}I=ܒZ1@Fa3A8 L 5z6L ڒ D[ @B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 q&dі z ř@YLx1=q:~Ob\AbN`|%.^ڄŏSe}uv|XMo੤s ἖Yvk\AWWh =5k{.e 5Zv獋H#^YZ;9eS1a[޺@%DncorbRHZI!D"jY@/?G3 {ӧZrWmx.*/aj4.ZX,2&ea)}`*Pz%ՈN'YTʠ:dZZ#55/k޸^cZ4Ѱ<"l, mԩf KZKl2hIڹIfIh׿q=26ٯDIz= L0޹,+o){}q25k9jH:@G hnOc펑tz<=iĆ_?[H.f ̧rmT:V*wo~iƣmE?Md> LBn&73|'C!/ҿ߉Kbt19o;$r;Q 42_%GXfTl8U\N9}|,rR!TY65O¬,`{RLnnݣ0guWRl-@GD((q5iT2{Dfao!_9vîY2[DϴExM#RD1 M:Fj&7SrƵM\Hvдn7l Qal>'\u-\PGuRYhѲ Kkw=*sЄ%&,/c~tmEB>ඖ<ejDޠf rIwrÌtXҐCBgKA2VJ_ԿV7,en=r&ߩv <Zv8dW34FD,2CL(EE Y XMIQ|㴜MRn5 іG ޱޯHEk&Sì}JLU=O Hv'$4nWL?ë5mHIJkWǖj:Y*-P ʻ؅YqQp*}b$}ݼ: 5T5jJG6l>>$|LPxkx Op%Rb$,dz&Ug( u7&T6&r흊[(dr.`Û?F>V|p0v2H'lr@- 0}&[zF˲?χ0N̡ŏyr\7U}Yy48e~t׷q}4@^&ww=oW_,fx<ь;7ӝĶ "o*v[uUcR6M~_}y#x=BU;EGԋD74 v@05]oNj<' {r3UWt9n6GMy;HʕV49ǎ{ZUs?bHy]z[%oy7lm:(?2 [FF}pAUL&i'Sx2-+R!SiU?Yk؅Crӗ|MJjrdS,rfuHy\Sb7pCJ9 B*ۿ5ƽr9HF5=0Ti7;0f9(3(OϦ};GF)-) =xxbNϪ!8::zSM^&>Ce4'90ٳbc= >5 ]{R ZKgaHq]Im#j=aT Ny_SzV51y\Jsq&i{32Qq6>Ǘt|1\U7o;B=~F#hZ k-)7y|o_̯P!/^{ëV oٍ%m~>y>!f el^RY]Y982P 嵳fْVt`v(cj%a<ߩ9Kk/k%T)ZJ6ڗ{- _*4Jk.yj|&= jxy&οAu ,u=dK_>읬6!gTt[ўoJϹ!jΑB0Z ΃A;wöV_^ĖenRweo^9~f tP}i۞ޛeݙ(nt|{ k1;EgaiVeeSTo,6fml,SSbL=Nq0xŢW( Udd9BAR|EonfwNhSg+`.xՉqmi.gi:Z\I#|fdsws ?CL-ao1"7Yӗ^ ^05d;YҚIweSaEdY ǞۇU\腠E>\"TwO\+n,M50Kmrn22MJp!:  N߳k } sL^g[(=+;P31-6Z^.ܭv|[j/9amwf(wԪr\FNyC]RYsZ:s]E! q3^Cv%/ME5a1'¢.oߓa4E0dW9'xVl=3M__k Etr̰]UO'N/@+lZeXmhIMg,ØXp:1MP2f&k6fӖlFJOjθ]O<C M90I:1і( K}ymKB>eNqǣ;y4sJVf֝k,3&7cMhܒ&EMH<M`_qK O3`K0w3ԃDISEr._Dz-ٔ)Ҷ\BwrsKfRbXT_ibEdNN/QG2yv]iS@5C MRE:\6=iR&p#or~￯QT̆_Y}Op;E5W˧0Kސ6A$mC7?̼֬X|BxlvϷ)gU>=O%;}dyd_bD"D޼F w!(1#i -LCkX!$V¨^=;|X3$Ȑ4E&Rrh/LuNE U) %JCۉ|Ȳ-h!kc30c <[P-]?8q;er)SzuNpnk? {GbݖU//⸔ ]2wbkwБښ tmMi6NbV[|ϥW) :E,MnfEۻ22FG_tpbd㟅OfHAF6dn'%{YFrvnQ L?^6Щۊs@2+}: b[`K5_0_pY\\L+I-sPM 5sqF&i=O"!BPT_ՃEVL!k{8<"S/x].ډIꉓYxY>"cθF_߿-(fwv0(0 GRy`sU& ǔOn@)lg\Cp`Z֫C 3.,KP޻nu.H{9x)ĂZ&M~^rՒs8R~({ 9kOҵiLXCbӻNJ6!xp`p.mG؅D^\oxm&LUƒDW҇D5~LX ǨUWpJ#9uT^dȖo7}[Չ\ ׂ3,9/dHY2q4JG Co)_ܥ5"ʃ-zK6=N| Df0C$g$AK ɐgA/.bI">P"D]=B8NtF3ɽ(~I3.Igu| =B$7:qF# 7*i =}Ȁ6GT⸐qzp}s#FI;/ȍ^Jqj\ #2|J,9L+qŖjd>Npb$ K_ȕl4qC ,Db}{^a;0H!T-F@)[lx'.b\8}S˻Ỏ$g_I/7dp3TIoLMo}J$J) xh)on҅#1Oa˹G CR%*+{r[r{8,먱^TO乍# %\HG, km_XBK}mc:Q˩X8}VSsK _ŧ% +H}* }b-cd S IT# $=ˊ)hf(DҞk_qx7?=so9QH03(ʓx9^^a,~rkH- yc&e"Dh_eW瓗2! "N$cw Ü qr0&EN߻d:ϢힱL0onkNJ]FlNu'yozdVz6̦=- _˹P-ϥW_?Rn''O^+NGOD'uO/RD;*P3 @aD*0Vq^\Z_LRyRVVϔw*DZEG̈PN;Zhd3h c&0Q2c@),e,Tՠ+X;jyĄ|R, ;2e@~8(Dռ5lFnop ~V1$?RHA (11 c*ECPbFl6?0@QPеeGn9Fg$Kq-ҍrkq RnT)*H5J:JalTfsdJ@7ҧoY!7%jTt}PWφH'ҽ VݑK.f"K4)H"eqXe B.2n~Zx\5xmؔCj0Ng`FL`&2#h1LKd2 =vmj1o#lzXbjUfsǘT/wJ< Fc< 5 $D*&DOT)A+OYՈXX J.8Q~Ymv[mAn,6՗ee툹k1ÃwcO*T^W/ݿqajaDqZ%³D&攉YҩJQI+Z?б-ŴF@q:}pA˓>hO?alwPmb0-s <#/ w hؗWJB1Qhz81Z:jASN6S*7+$l<ң84Kzޡ2++%Tv-t=mNB(Pca©P@`*>8/"睁en>)~!!BFAG S *q! ۓRUCff$eD ­n%m JdqOI"t6}Eyl 6 snDDDetf]vE*`xbyuUǫA&A1:VW{US>[*2ǵx)>g걽1t`~:G)ZՅe_M%i֪?R N](V ,Iy<^ED%!wRzqB'Is"Q @FfeOz4b޽SWbK ݣdn>FZ"ޙkgaꤎTgܱ=aaKa@P]=J72nٙ8IM]=AZ$p}uϽFvFktL}T5e^hatF۪2N︃[jXctc|޶~VsۣvRԘ)jnFֲ#OG"~OHȄg '늓Z#tW} b$nK{~Ez7{ZX$*fIB=8qdtG⹸[DdTI.hExy{u#Sg[`5S12oo ƅ@ 37yrHo{[>kN jypjTCv^.$B'%BSRMPzyHz.r{qj0dq/HT r 1ڂG_Փ"]&㘛~ޝķte $BRdzGБ&aJK"˧%2ƅ| 7jD!>C0\\m%+pL6 Cz3(=! Fd%nsiJ4<7ŗ":?ORXz6-y//|3$zE'ZNGbCJ@P)E"Qѓn¤ߠG*Kf26ibrkrۇ%X _PpY`xfHd5$Ib hcÎj/J5b潢MU)+V|A:U؎1zU_̵qnw5~X^8h>\dkESQv6qJjrE6pYIR>zea:NԴC`09NJA$e# $7G$椈4@N"i-*lHX̻7Gb{;k*܀T,Iy<^2}ƥm;@ f:u>W+Mz5 0T}2y E=k=?Z(" c-RhFb{Hy˭ֈk ̃:̏XhLz$EhSb˷͂ aKx8޵6b pa9dI`7wA7ۖ՚ƇǶd%H䖂ٱg,Wzq-2җ9Z&,@-+VcE*̤-/\JM`"m00D+0`뉽]OySB#l1e&]14H؎>w}XKa\P'ޱVXfzL] .Z<8ʆ5JsЖ״-la-=-NW] L k8$!B-yd+FB,άhjf!(-kMmgr99$F?>߿Y"(Js>6u.#e5(&zUj^~%MUtM9K"%ϯ@؉#{R^>8O0bќ"#$.~( j X^ݶ:sjԳ8Z˗|g?C΀noU ~ōN!צn}bՈ1{Pܩ|a&CN }\Uh,hGBV*(#o.Sޱ"X.谅3&հ%O1cI&m-Am#2G{cm<ж2jDh?Mt)5pʌnNXP4F1^ϻ|4WW8+>1)ȳ`/<8Ž;jxaF Bkk)]+( ­0钇"yH8TZcbkh;>0)h`>\Fp@{\̯5ktw z BHDTwkV 3Hr>"G3~"$r1H%榴lFߪk?ADʲUjػZm!v谛 25H"xb]ӽGqt Nd\:dRTS˳&\~*_4~ ?_M1n&i줧cnPg?>Zn~Ppfk_%H*B*"dGI.%q;_GQ{*~>jspD(zb̃g]\3qcy! ?\;q`L$u힜gDTKT:S5A$k㵮9<}"+9R;eC^N $q$FmO8,kֵ?xRd˂-b2F0M'yꭝVķW|n#sL&ӛ?~馜|KEF/cm՚ ė;qwK# o{" 6& :\Cz%jd}{AxoD?V\)c,+/?1'?A65?),'Q=z Yx!=΀+*^XvpK}Ƀ&7e2+u2Iso +nef?L!GFc#}Yo~g//eu^-ߦ-0WbLhxf 1>Ee)HgFuTW-9 ԰fa@P*hPC8$ ,2k'@V1N9x~>^n4`ʛtO~9t*|]J]Q`@PT]),.r (vTbDU8.pB ,Vb Τx(R&=)ᢳn[j^"2¬~M ""U){HNOQE o4 ?RH&]QnofJdG5D)tzbw'.D2&zK[vedl;)Mfr**su p'%>bJAd}bl1 уH.J2b*ta䤪!JIIUI9oh2/ܺTq^ԏ߳dTDJy!v s$ ^nCE1Q-% ųC"S#I%,ffY71R}TRq#SC@Az{RjؠD&N|qB(cE~ib?0{#)۱ʆ39MS(F GJoj(@ =4U]+I#8 П ܀V8cYeV|NmazL$T'Hr`xAD gm^"'!߶6;[vL3R-S`\y1 jj|a7+ gR&*սQ(4x8Ih a~9bY s1H5H4ܭ̘۫/ӆt̼## 1x5NZ]Gb[2`#@Ȭ YfJLև7?gAsf26Hq<ξ?9Ujyc,/Ky1.!oH )K(fOc\ -ʴJ *1P1שu5x׻8V-ZR>L4"[7 Yһ ;g{'(VX46scL'91`UAG< c4ckƸƗlaZyǸ|8GԶvHϩJV vm<}9X$>wjDFXpԇ/Fv$a0'?'-:qkg EyGpmP5D,f7+3(d<$8|PqqnЭ ƛx$z?,\e ܳQ!$5P؞ %sdmxeB0$H>C3ʫ; Cjr2I)KBcN)T.[ P&VQD=9Ef|:Lt~k]CǙC(!L Id8TP%8e^hFGsgo"VHwCSBn o)7]lJnnսbЋBۛ=@Eʔ2P/j$-383BvN( ) Dt&ÑR@H>>z)fƉAK26[LPBC(TQRP>șm)-I{+9Wp[BKpNjIՔ7NcT$JOL[erCfab Bn[@UFOkD!q_6HB!5q^ I6z\-a(@"d7rl^J>L(vs)/ЀΖ}7˪ q6/gI>k@P@|c_c3O (PF/,˝Ic*HqvM5e$ Q^L:uxĩ8Hq% "$~Wdv޿׼ӣ&;!zدFv©X!:Ewx 04e_v٫nǞv{˷bSȨBq8>X䄶>fPX?plqbg뵶"{ʈo$BƯV `TEFXF37jʄ$-C* +ooyϻJ/l2=f=xppBbCzݷ? k|-1&hs@4^T) Wk1ll:Uˬ6Ső3fz>uNr% 5 1"FA hn{S(2\8֪&OW4tA ѝ-5np=xPQNHdj-Z LDd1R!EX0t#а. 6c<OZb-^l 76[;nW>ܕ݃Z3q3DaQ㩣}6(;~G֊( ]Q49)2n5B!1LtS hX;+(1F2rZB׸m^̓38,Ǐ̐ ,Oqu`\T HL5Bh35j8 _ŦSc8,j6Lu D[\+L`!}RO Zj$lwj:ƙҠ>ڡukoYbλ`DJEHu-@T;0<T=NIixlg$:-ÈAҡ`m :G#~mv&SKŔ-[5bBIM}x.F*!K=릻D1TU#u׿?7w6֩~ϯ~vӧ~LsYsfV^C^k2OoA*Dc0EEsL J@d ab$C!|.%?ccg5rQ0,&ZB/ !rd2>% ŲՖe[۞Cm.*Z.hxq9uy5g^`VXsKu{s[_iݞҔ)@-㚫/XhQpr[HfY1}0#%B/dJBGl7ϱ£W':Їg員uh0EL*Y{Ob7x!MR._puyѺd4>HgѪy_Fa"蔸qڳw;XM6de፳ aeƯR'WS6`pXÁbOf/[.?%ۏn)%Oh?'6-A'n!(u| uC2i| քS~NVCxa:*~YDі/iwu10ҽhF9\XK"NT@m[-9Gh;2P^KY|/Wpqq@b3zv׮? [!-̠P i-˜%%+T!ch\:Xp?=rY﫻*[̗ېYN+xarTy U&__0ZBzmf,ʭed%#8, z czo Mm?Y`!=nnok1r”Ys0v(:orS v4Gal~3-&@nq<,rՅ:b{Qfou3.p=(j,:=9|l ƐV1%4΁i4idL*V_Wᴔ2b6.$ȫR52.ҥ7OhH(?`ͼps7 iUBZh/S>)V`>^ u3ЧPȘ\!P⫎DJ'0&(EyL]AxP[Uý7An $%_%a!}Wze58b:9FǞw^Y1N-͗|@yxrŪO Gi>xTb 7N<0v; ~wx$jpu%Ux9o-W5BI H|~:|Iqכʼd?BZcJjUZJS__xߛ$H~jGLx◹q(eU8'hUէ޵:葓1BxHp0s Ñ)Cq!`y(UNvkW;'4]rRm#nxV%'[|Y58bG,@޲]!}JXPBoy0#nǻ3Yϝ (D<1{@x,)s&֢ gZ{Fs D9ժJэԘezjT X#҄ YaICYaÎyYz߯xZWfVbHIb8;!QV8ԛY1%_t0O3"+ܽQK[!e>wWWur*:zQaIQrwe S1ݫ0tn}g>b3 7~ p.Mt#7q T!cCw8|(PGUzA]TKiǘ}\83?)1[:SjpSC49SfyTYiLBklѢe_jՇߗ`-谓@%ܑj>Dq]o#t!f4+a23D741' _ٯ' RB5f?,/^4-qb)+ .kd* ; 5~H^/$xO]iv@y6t-L?/OY)١W_p*/d6wnlva >ՠ7"4')ÅB~ܒ/Z9qZx*axHZ8IT91iL p'f(s%5Ԉկ7_ S[%v8DlA{:BJl@PVn!&(̕.EXA Ój:Bܝ$Zeeĝ&UVq { d ^3+$3A]ygut$XQ ս%RDinR3;OI 0Tdu1)Ks/P&UɺZXLKm)+w~. X "GEF`҃9o=2ύpDXm_{uy-Ax8{2:{c3=mHO7rWʸ{1A/^ 9o)gn]gV[Gʭ'ߋ ([HB* )\ { ^4NԼ\R"*qQ-^ ӿBcN-* A  iePaKi߻HY)$HQ)ץ7C3C;G>;ZZՇB Vc[ĻC3bkN hk fZ @e0Vx0azL qy8gpUiim٢YPwu\ 3 ||S eWq/i$#L|zMjċuMzfCL,.jhvIG_,T]EfԃaoՊ,JEѮgȎK#zLRQęMggjNG4?muq\bL-jKk;إ斘7RBY;?;<6T3ik v:kgӷhphc̷\bap '=%h[n` ˯LR+d AA碳hE\~w#11c{!:(ri L"q+ $׎Q*dT8w9&edZS`Ma%H(^"cYa}.QԮC ӛ{ujD-m=NI/;M2GYRJ{4lWxF?]KU<} 4_N18._0߰5TFuR %%ANיQ9όņ TƲQ̙٣kM͹7+4sQvȩoR^Tn#<ڎԤC/}ѐ7X}镪䅩.LdqRᡪA'JsrCAKNzlFSno.- ci?@uU ަ6fy楜A'" UÜ{y܁_Z0hzB{k!u%󛻢GDRjAWIY缍 Q1)45;To>^`>Vw5r#WNqT/V2YxxgSHL"M1$5M<#pI8p,UO?N778&p S>e/|3[a?s5.<߲eSOGCgKê&*dc0"[La 2Agj3oЗ0Y*q3I+u _!?ymy[ W-#T2)ueY& nsX{cu䲔?gbu2궷ݹ+h3c8R/vpzSJ/(RԿsǸ-wՂHPqunAmL^Al1DPQiW N}I^_{>PY=kb&n{.k$ ޺fغq@E8Sr]OIɚUB~lZ؉ֽRoP:d@$)N ^Itwm=$ J'DZa5=ZZfg6ئY?ƀ]MUe/C!b\i5Vh15V=NY4,Qy-x&=vX!š™2 AߺFbVu*AI( BxIq_fWӐ`3̎鬠G*5'⦭Һm?Zyh[fF`F0>DqKj/9C `T[>sgӡdIAoJaG[:jhJ/?4ELѹo$ZQ=>ޑŵ**W^Fs̍YH+Ow8]\&J#w ]9뾻8Qhമ5V{Lr:_ 6Ḯ1@u# ĹfgQs,Pl9 vG* <'F+s1{L;~]MnpRԒAw/ަ7_{쾘~ 5LLؼ> )+s-j .$iTۼƥаp6_KU g̢l܇~/ [z%HqJ-3 ART!ǘ'E")0^pN*hTT]Z_*VCM]$aWP5ZG4>$539kĶ> O*cGA &8_Ҹ;%a6ŏ7nVrկH](a=Ʊ?f}m>lk1!{i9Mmy+*rN.V2hYqoNKkM^]߿b eO_5ftVRbOowu3 :0TfAɐLXC E Ӌll";o;HbCRNC?u1{iu*(@[Yڴ/ +"k2C慢0]ض {Cܷݵ=Re+C=[l Tt>vcF0T}]ꉵJk0K\LtZc/'Irյ3$qvXɿyRCV؃`H+ `(\c2yKZS=MOV&j8AqFPJxF_=Qc-2 )^V2hogztLMZG礪Qr8pQd"RiI&*,l \x2#e_V_*;=0+dWǓIunR}n*)o٭w7ˈS3oLڭe y+e2%i6,%٤ʉ*z}`N\]$k G7>QM)fE' oM\\Yp?pE[iWrO|r0Jm3C- R {< MO:v3Wf44my]_ Nhx$с !n.\x\Be V+2~]Zzo!rҪթ3o9ڔ zMO:PgLo& 8eDA9TC /f3'@Y(Kͯfg#utD̾\?ϓ\X;jBв2zjGM>y]BJ >1Rlqa0,HP<>@1c:O'TyL:j$)# bHLTqmt[FE/i<zi @2h2cCy䨒tS=2xtMzxErв$?u  pW2fr:k[K)} KKĽa,H4 T!-Q@FLOj6 2BzjS"#"䠢@lNdF翉V#|]Whe2;Aֹ oBtճ~ǚ(o y4uǹF40\7Fq>C{2ZflZe "[^7gfߥ{8fv]1pa863PGz )Iw gD[ Yf-Ws$,A"]Q561r%麧L]8@V=WJ63{m|tYg^t J9H{&4>_'ѽ{o7qUJ+h)m,S"'X +yKK;OGj'VA]LYӷ5 n 4kUnݫjK!'ZQ'J}ĥlz̋ tCՐh'd-!ۓ}XM_ }C?Զ'3D=ѸW:߿~>d1n&\\W"VU.t߇}$Oh;D4 P2Hnuqb"ŕjCZ ec@-gs6dT/8pqgb1)/p*y홓!(yt0hJ%}LMyN TLj:w||XUoW*Bxee:'"TDjt:;hɊkFX)LtQkȨ;L,SZJW(?ǔ!kLf*JDa .sRP.4ONF (-yOe6LN)E3U+mPy-"WZg<*gɵ1jiwɫ>Zg+hqYi=FsziG :+hy'ggt8rq?Z:hoKLZm ]:rk2=#JMXvga)v x f! f%QGJMh.ej=|@J ))A)gDłב)R>pX s|%$|Fn?]tL[äǙYZe4Y6Z9aZĜ_Vgkf 3];(c\1^ %DGk!3;vS('B`1Mb(Ii"MR^mM*Jm͜X0ZfpJ%(3T)alyjDjƬgu6$)gxinjNmW_irf6;,jv|[^'٤<_݊_%0}5MW9t>>1ZZh~p&7xF09X¨xPLSfaLE~܆ @YhY,.E1[*G ^|^IH՗k]P KN,-`ʼ: Y?uj)l-J{Wb}kA~DFJR)X5JECK[59CP)Meb˻;T,e ^uh\usYhDkCNaM*k"ue.ci(3.J3ܜJ w $U4r|'D< 2||4)D7fP@+8+ABl fqІJź sw?bqp|' 2Lѐ @9o.x x/)lz#4h$̦yˤ\(gJʡ&_Ut >F[F;\r g"$GBF;XJp9iTFsVC Ua&䂒ٷh=KD,2M1 "`3) ƔlL$p RKu%5 "!a"W;ؒZ~g;I.ݨ*/ {b.l.V@g&d|6ɷᱠܨDp͓i2w^ƪNHeR!6YRƕRbbXj{Y}z.ʐ7osb IEL:/98RsꄠCqrdq66ƩF<+I޴\ΤqXA 4"R hR<4ZhA 9ik9),"v96 >Jxla ܔnA8*&q4TdR<޵q#"6YdU,WGVOqtʹuKXᴺdU(&M-6g)p'3J݅S{k;3QnyΜ|v`VFNܞ\"4~=S؈)"?eGCQndjY /QbŴHo1>[#ֳj/T^fS9nNyUIj|IYhG&sVjbFܜF&AbD5Fh+ɦ@T iM_F H3vI\Rl3K}MLvm:NazߊH{T;.^(n(k?'ʕ,kQ^G4%@y~VZ> ƨ~LRN]PR죗sC%D2y4=v}C50F*[a .ekJ>hjגaSr%A ͓ylU )nْ4e)z˵ =xTyдU9 d e1VJEst@bQZ@^ME&b.2R?%a:؎ɋ|$ fkuشlgJ VAqDeՆ0F`+@qHVJuS Z9ufϊTd@`l/>œ>7/'sc>x|mI#o=qG 0 .zu.>tTҺ}9qb ![r=7p61#D4??ns_QC#Y_҃#Ev%8Bޟ8[>N" ,aQYD*e+o1TKC5C+ |Rc,T9:P:UfcW|պb`9L{d4ZsJ{u˽X٤Bd T}vbgz|*O\`]͡HEp 5&/yq<~¿xz',p_PC 5sN8 TeCsWK [bۛWڭ$D=zؚ /afp󤭝Nu~>B{.!\>'D2Ba+$xo}qaf {W ̞R "H  lhf3{o^i?Qfːo?&śxPj_#{ўdu~7v-[mH^$".P b](I,r095BY8 /᜙ݩeWꀿ+\dq(r$q열/F].8&\3/w] ~pdWƜ[yG{vs_$;"_EB5jgO?.mwBߏ2ƽ]̘0PW#&^}9~lKI\Ӈ4Ћ1k xP`cX;So nj77nTz[k"Sg6z&JX)fk`oWcv|]m$7ُ0E~1m? sY! >@x-;7Cr?g]y҃‚(utLOc&\y20@,` asg$op& QHEBo= T 0؄HA8f~ OgݑiMт%Vn樼rtgWTKw;;뽓)0]Yxƈ1 v2ݫ\ pW溴A {Yކ٦}yQ`;U:vj _d.~V kXq 41 ފ[q6po44~.Ue'+r58Nho 6b;:N%ex6vD|ErDfi]5ߞŐwj~O}7GEą~7l/;P?s /߿3-;wҞ|pq7#AfY-Wed<{Y:BnI-Ywi1OG3$(ΚJYo58m)2db߈[CXI&,R5j7ws``܃b˹}> ά4#~4NjC:8}f]݄EnLߵ%#')/c5Z M䠞?nW)ާQeO'ke[oUt9*W#.򝮜ECڪQMͦrd :B,AtBBT0e f_$ ~5~3ߕu5c4jXLl,~-Ɂ%z_荌 /;an]].-z/BrX+tԹV[VK-{8xp<L3E1 \bܻLmljpzEiWvdCBRh $I)LsdlP:2UӬԌ-\~2kr ` ]dKcpBbZ*EyPNؒ,CtݹtP€1"bޝgA`V*m^n 1?_2.W1*ypmiazkf14;F`P9Sx0Ŷj|oJ,b?df؛#Ao`rHGJb&*j!\Y;rEncΥn$\`zNU8φi#1()`*O0ݱC2{&~yn%+cє Zf<`KfލhIn^3م^ߦs{Xl  9=T܋h;zTͷ'AjII%.\6*Rs8kWRr%&\wzU.U`MCH9ht& ZiɄIq&mvýx;"Z(zmWHxG`5Bړ6hXv+1ŲP5lZpf{ ߧhSO[W@u?qᆦoD1Z2Qjp'“ceB41|J#VhvNR }ݱoM`N:"d`Gbyl/UAͽ?O2C[//a<]u"V95dM/咅lvvGã {{hiw1OD 4tPڕ]Kr=@ &WnFxJ^%1KjO9hu&=O)k[4XE(] nA!bpt;.[e˷e4 Շ+a?w1O4ܒ/ynג&ub3"bb]>#H18dv!0"z҈\D6~LdρTm?"=$+b́A[4;OvV,O%3a3-CS&4^f]0y.Q[=w$v !Ӡ6\ &C7|ƤCJldCu(H=85aȧ}鴷 cܬRĮx!zK E@I V I-N&>ȨjDA 4@Uiqd[jBCSv>Ĉ Ss mkĜ|*P:LnfSizkmkI"%@VS$/y C,kDyq%%Sܔǖaؔ澬ZU]KKJz—o+}T KHǬ:^@^q1RwQB]kByLtsJU_B9Q1u6A!L`Mll ROy bLڙEJKU;Lpw2FzGN>AcՕXsDnn&NK[ z>{eafXܳ뭉=S+k[g { vfKSfY:|e3$#v.Kśe3jeHrp- U-N:/6HEGbu),dpơu$F@>cEi񎈿9ê<~U]Ju՟{:浸n,ڿ6kWK~ h~m]7I:kMʹy49?CXhmIWκi&馩70ٱ$?-}DDݽMyZ\.^lo7 %zu2u:*aץ)ձI!4& ~W9@8ޯ*D=9*i}ۿx3CcDB: f+ \FY#[?|l`+UC;_ `/~σV5V_}~\m;c M 8u~>o[:x_=͔y fװ_[_^oMkTUn|E+׉.֚` sk_^\^}@nˋyٳ7vVd6cJlk/uGIQx"?ڳU1yGF9GΏZ6H"gv(CכߛP^d U&f¯1b̯ T41mM:##k3*ߢ4L*OX] z3SݢqsB0"s' 1'M]:֋.z24Smo.0L~rj7zN. ypvfv3[?4lz^N>PΞ"Zkzs!ҪCh]]ߩd.dRi42Ll zjR8lԚ4Tk;[UkA`,)rK0"\"%6C%Wչ“0x+ fmF?nxB<@ -nIV2FUt]K(:iU7tkfJn:Jߟ[O//_>6+Ѣ_ ~N.WN[0y#MlI`E윚ۍۃ,\6r6[ :6c>uѨ7xT՗6x(*KJd:]ZFgЗDV5=uJ&/Uoɗg+#)M3Dqq_#R.lÔ]%ʤ^ .Vmr8jk)ɐ%hp-W03ZՏud>࢚`h1G\p.+]zӾECu^hLbqݦT !1@"%D:J]G#R'Og0 :)k<:)cI[0=۠c+"kH޾fFHpӢ08R/n%G9C…UƊ] MLa#R1`Уj: DW*_%596^2A'1)1 +z2'"ZhEqb>,@Lbͷ,AxbAr59 'l=úue%F8l݈U6)粨ϙk+-TrY8]e"1 4gM ւic TH\S- w;kz_'w;Vt;T&E3čYQ4U?r~Yard=z`LXas$-2ג# .rG?'bCA&4 :0lbx!4X;&Kdm.z8==>l! 6Is@1 Olv^WkRƘC1\L /hCM6V K 1 0%j-dRQXWcq1='Y,zw,OQ$>>.1{t$i̪DU !m;ygv2e{t1: ;u6ҍ)DvTugkKg0` RG ȏmE] јMK}9hTAK=mcV]8;ݩ*񪡟 hFC?o"CNKWB}dq7P׊^TO )iUdDb"XN[6^P:F;vY_&ql)Sd> [j?i3,!羽;=9xd>O9OǓ@a)ȖYP9h Mk$$>ܒs=.0:tjވ_(? b7|CMe@kحAzE3ӡc)`s6KHk_(kƁMOSKᜡ嬵&ߚxoJ7%^R%9]C0uS|R&11[ļs(>L9!,CN.>-^^Rn{6s+$⯟)$ks=$QF!.e-{ckO ]zR5 srXnbUHMMBv-(!|cwk<=}ꜧîזe[#59cŔY?r;..uyә]1lxLY]+$o_sptO}8;AkצP1^ 7;堒+ -űOB6`niaPfC/dVfŶ[qq4!9`a=],]597Kl9 ̔ΤĴ؁ dGc۝d^0fH3c*I\ >})L\lY`Ǔ 1&5 8M疅kh=K6S1F\y쑰K^|tlָxtW%N5,r9뽤4aQUwP<;g%) cCΦޭr{6)$@]*ˈ/\Rz I08l/1Ћ|b =/R^zT"Uiăj54;XK+djNTq%A\ &!<^W̱h:l(f]UwAǬgQFlb UdUbj j_x_Hw`Xl1BC#O1IJ18 N4`4`O9wXbH:Vgjܸy׻IEQE}pw˃Hfo:i?kgw챽^mZyDr)ĦF'ѲrT3AymmxUZu'n|(<93%!wBRe Jviwأ[J`ן_/*ED_W_/ )KJ(Wd))!71 ^@iJ:br"m ~BVТ<0=RULw7mmIFꋹ+Y1!A {WyBD-8$JEM3K=)֪Q dgo]Ɏ!}^:[j"8/l Es18*!Gk QUayHRR0ZZ*>~lk& J8R"qLB .2#xL)+/(SNˆuYiVv, A(<V+׵^Z6Tl$=w%6^; ;G#Vdt1ZեR\$(q8Kp[S^r c @.k璻vkvFL­ =}VM璷;/z[\uu9BbVR⧙ x?lŌ]~0[=z_z&/u1*ُe:l!LX^/rX*  U/Di].Of[WA,.K2:n2$^UW2qR5\<IOٙv,Rg~K)ar%f7`5gY)C 1.(ٱ' :*y>*#;}8kgy_Ӗ2(0)|$ʖXF~wq[c(>^ ҟeR9Dޙ(U Ю qپyA=MAJ{tj? .?^?uKE*ub xx߭DW(ɿ@aii@c1PA1Zh)R/8TE1]i;Wj:vرAB$DF\~;Ψ/ Ma}?֚3qFP/"+fÂQ\v-Jiz`Kj|YTl<`nN+$X5#XX 2 n-W$WACg;l0+H6t PRzءcFbC;;K~Empxy-@}݂Œ $M5_s>A]0h=Rru҇Ū1>{,CLa٧@Q4tV,iA ~Ͼ6M/K\SZ3לdd,u22IN춶ӱ`;'oul`^qUDJ\EۋW}D.]&yϷ_*t5Mwn('p}#c{c(=}FJO^Ӡ>'\ӝY/zhdUUM 9cU)־,IzE;v3v~i8&N%0! c`)rMi1m \+wRouЯTZYkBc jU'͢1CB_&eNQ:x5v5) -0@x`x+ iR9&9UO}iajJՓ$i<^ym}z("jPdF 4_%Eɋ@UJFdbL:v-ܭuGHqld f?Do* ăaP`GuޭwDvpk4z].n-ܑ7OS0 p񛠴\.;Y NԶj#ZXacK/h-)NS>XEM"6kh͚}H$`z#ʉ Q<;T8UtJ[?}PRl]ñy}`t& WlV\DU3FF'Q9qtшL(/Ta@ =2(YN%Eءa,B[Ee@lWͮuS{c6˘ҭhYZo KX5ߍvjI`,cD؃O=X?q+-Ĉxꍽ1ʃ -]:rcc#d 7*x]֟8q숩_bK{>3P{Bo6>9B&"z]~(hu|dAɎ==AI0\-Td4@ tXh;ŽvMBBCǒ!zFyA')Tm6lVZsfs$Kw\v"0vߧ{)ٱkMϺ5bկ&wiEQfkG : lr4d9=zj6c lѨ7H"/I˯SF=v[y{!-TFi|*ð=ϧj UW@\u6h˷\ޜ[$vdTCRw< -GdI!7`SAvX悎w Kj UXӥs#]~3u?B 턍ο*GvVquGh7B j_eÌOSF!O9 Gh7Bn浇YD=f*!}T}ѻ^a*Pt ų[[xȓ4ʑ;򵊡"y!v Z~nzx[Y%4OӦj霨}~E~ˇs#D\C%CCge] uY|NKR0Kc3ZY8~gk*q{O‡אx.ϲrBkN2n5ޗ`VE'5S1d!1iXϧgWN ~br>dP'c;g\L+K;X%E"~[9o?} Otpy2TP+0GJ&,Z4PTˏSt&noD^,LjI!CMN[Syd:x`'vV:e`[#NnoQP_Ԣv`wY!+ٻ޶W}y<([ a0ERE)iZE,[ĶyΜ9~QB!<6̿W:]1[w\(U'ANl<.`Z]5X ;Q>༧Jj-8gJ#=}rA}GT"YVEgW/}{-j^p=eV8~qa=|Tl%B*v"QDvk vLbn+*+YS.8bGqBl{3Ӂv|Ut)vWE?c;<ܯ\+{XbnA4?t;$Za K7Ǻ 3*!KJW`JŸ~V#וsrF 7J)Fx埪Q:GQeMҁu'-n8 #l$-fs\۱ &@,SPY\ khة8n 6S  4 l;=#x0JTrI@۫o,@U2 Fg-b$ bQ!&\I!fM,i{pGV#@Rj RhCn4n>+U!AL;+VDp(;Buܳ_ {@9 vx#ݷ_~qGؗC_}CseUl_Qx<^Eqn*xi\mK0Af!F%ubL}{+DdmWG7uV@l*Rxg/_Q9mA0⯢HÌ>2.LEQH0{F>stk^Z*JaI BBN8jb& kLƘG##[E N *-~z`YڴX Yf|O"$Qȴ"@g@HM?K6%XU,',4SE"Es`$Ug#7U쏎W䛿*t'MS0g,X$DZGDC(:213Ą#-%++Tpw&%s߂(xe'ǣIh@jb R?jyOqW+O&#Pa÷9mH}p`1si59|>_]a#Xu}a8ʹi*=Ǡ᫙&)&u%bYיWⲐG7u~Xԓ/ A5c`ns(j,7OK#x@y*\bHy2ʀ3p4Zo.Ξʼ0;o<,?߾}?>iJ]nlOR7i$H@Nz;pxkJMb<|( G3pC>*|@1=|mdcs6R!X`wJSk-t{;$O;-dAn+gdE*0)rOdq8 ~g>1H袲(v~s\A$:h(&N . >t[/h$쏸d(8pPE5/vR1LQ _ Ndw?ޛ|)3h&/ O{6q;1q~Ta +8ۀ#P{~,`KL&Vc[h5(6&Ef8Xقǣ r'P;Qa qeKf _a ZNFv5n8?>NLUDB!Zrhk 1&!FZ6Q&*nmZ?npܪRB{3Zg0L8RBnoΒhP!G9hZu{"_UY3X3rK/Ā}vM&x<>Q2 k+JdjΏ͏I xhH(ral=; 1ž)󞣂ocA.:Y q' O~8qD=ԏPvV,t3m%A[IӔROrF':Nt`TsT!x%+t '߲S5(FF;诮$9#jޤ_lWJT}Qz{߅!vwV/s)C[GsBR}&;#Fܲd mA(?%z>Չ2֭wܪA*\t2Dyks[.hٹT4pz.ٹ 66kL*Svebb2 `V*Na5}d5Δ>Jq<|pڠՉw1L Mzf62kVwtƔc'=y[t&UEfs|ۑ}$VEG`zٝ8\giͤ20?G˔BoLxlvw_{{k1st1W)Ƽ}ovL / E!:`>_K x~a$g]- ̱C\8n0T->ؒ0USνms3|1KJ; JJ*Y1 bbj fzxEvs60OPB;I>1{muù`ebN0a)5 Nms>ih~8M ޷6%P{CiXLmZF!9ԓdU#uӯ\w>0-ë}pGnO Z1R9-{_DnŽÒ rg~ݯ̀_)7O^>US^JW;( Ra4,CWn [c"1-5ҡ3DJйȆkτ*34Aa1ajC0L;HđRQ!F*tU~9#+tMr m{urPVUdj. ρѿ6X"4&T$OY!"0LR#2,rsp+0" z[@R0b*9xj,XpN+Z="jM(`  Bqr1 #Ju#I7h"ih "D$=(~!!eB.$o}Dmlg>rGSFljfOw$. ǟƾf|r$?ᓷwuwq Nz g (ͣ~tSf 헚l;y3_*F @ L9炡k-mFֲ8nkR\U- i(WvkiooeS-IOmXs"F L! E(bk 2-rWVJ\IopZkc5Ys$^o1M(b-E8gk[`%` %muw|ȽS]^P7fLPO,Pf{>(0ly>۱םlp'+,J}+!`7iG;lbzy Pfu;a>༧JrF&6&5'i_\軹/l:H$lyb%tc!h@2X9+wAtUfUኋ#bC 6,2{Fn$/ m20l%$]ϱn4ciA[ݶqK-Kqc,Hփb=M "kkb~&#t<=IBrXX +(ܸzL,2H5\X?2~Ä1SQx,2r^.! @S Sr[T)g,Y)l˿}p]u7g%X߳zoiSLdH!*2Zv)[ʜS1ڀ4Ske1) 4*'7mj;QMm'ߚS5inG2 @/Blejt (%rPS<ˢon(i/ɼknhQ/WUeТN;-趔|%~0-aON/M&_ #ZT73\ɡķ!9 +2Ipo`LʶZzA>LEp"bbdKATR5䪚%Z8AH>nbzwA7Y͘%Y1|^ܤҺ߹g y`9ZO2!5q: Mo }~ShUᇟ_9닷l{_^E(.wd#pLߪrKøu,mV-KW]u\{=v:m~܇wUHkpE : tX+P<矯y C'n5X.j'FmzOןBGdbՑQb9 ]T %^.E]g׮QzYMGVXbXkvPKҼ6u/-rf8uPE2Sەk_ME8P-Bb+pí{ӉP=Yxq K3#!2% 1gK%k渭T)w|o0JS^7h;h6cmD̛A%!n.]nHCChX7oD>?2&qکf]r5^WNϬF*&1en2#JJ h"tmFk89B_V_~aYV]OˈW>W_^DRf,Z6nz9Il;Mo[.E#b5ހiSs>,w$ "1R"33F,b>p䛙X߷ܳ5}:̓ ,'&< c6p雕A5w$<e5&QRG&A+jP{QfܪL)RhbgT!cd&`xIӗl]}YM[P} PDFJܕh⧶ZVFO?}y~ˣஜOƳOU,4͎^_]#KmP5>SѺ548GGe[ 3wl V$Y`[ S !L2 )d <V\z7 ;[oލh8 p'D"G ^~EUEvc}_ߌN7,ݟJ7WF߸O i (O-:mde~3#gi^q-Jy !FFnR2S! _'em["ctv97nR>n̽0_wNiR|2k7ZcJ0کaTXm( i߸hwUb<=OsVm1XB iND\2{eك~<aJd"Ku x\QqN3,ח1mP N?VZ~5clH08H. Lu.(ūϑǐq"+ x (Ȃg@-,xR8`]403Fc0,"R"GbJɤ^--^QS{QNȲQ-*ш&eYtJ3;5L 3xtk4>H.i 1TȀ*E]%2 o~A0]{x xP!4S@ SSZE JOn!L|׮PK ݥ2 ʾ|a[@$暈jJ}tܕ[b]\x6NS ":#rWH|,Rj {P&}=F+u~Ś N._W*]:&YzxŇ5  qZ*{2e-i^SPQ 62PalN`]نs qn{N$QJZ!zAݘޫ>JEAFˠh~Wmf(##%^ntWm06%?Zl@Ո4"SVXzވ@kZ^wh%`_Jon9ֽ{bv\OB G_C ݑl86z|%=I&$hӳlHPD 1Zom vEB񁩥7PVIP.[86w3k*:{z9)yo!iU佇>,[HͲ7f.Sˆ.zbjG≕gⳆgķilӓF93SZ)]NMf3U*j/P5)ѪE?\ڠ*OVrCq% 2EīH9 "h3^ͱCµ&i(/~Y~ңJJqw<[o$52ƇzV|gŇzV|gubp=$,8>(睳&)H'y"KYc!4:wU%OG//DX]gۉMHRTCU;j6<2<ڤ1<'~/"69؝$@6P$_ &㫬!h!"jZ :''-mc˴#7;x?A9Z(G(@TМ"PI(TK`5樌,>$"B:f(~%g#Ff. t(lT&iM6JT٠w\`噉gGɅ%nU9P| t"찤ΆlRX >UJhVk% +.L xX|Erg6dҞ묽?=8*g5bø 5|?o3Q,}L{^;2:5ٵ]ՊU|sdd*{CZ5U;J[]BgǃE^ ^q j^7{7;_=G of oa 8M48")¹RN?}5Vz:KOU(s7w>6"њ9PZmD#m_QCR{UV9F/4WH 7ɒ7DZ>\^,:4VE(oM̶&ngoE+av9TLֲ mGfzVL?@RB[XsTXJ>O TCV$ wzFK:Ѡl힐J˟8+{*m}4WCj(*)OQT2NwtaҜ ł(7'RHP5$X˝JSd|ҴPK)$I7k0EW8LOMގvv9@Dt .\ c5pӌ|ߛL%3Bxg1aj~h6mQݝ#l9/0ԫT/D=oi KדJut^r lzfyE@% avJ{w}2'EU?i2bED }BixEԴŶᨳBM-%W0!)"QЂHqm`V;c e%'!qO^~n4"њ9Pw G#41XA,5L':X"o(:AUٟ@g&u {}vX7YQ@ֹ hL +i=oF Q} w QV6PC!p(`l! aA=zxiH)4l:\^0?L" d , o7 >)xҚM!&(RIdVFjw1 齦)2SBzEmN9B*I2 ܛB*G+DM{aE 9H 5rE^u**85sQC WBtPۣkUuS3aut[{m0%l<쌮T!fU cr3- aߨ\}aOYND6<5S2{O 0$(HLЪa(v%KXm!PmzO8/ oYVyD.4Ǒq.XQ/w^8Q)wMnHek1uT^Ω$[Wk ȱi"i{}  RgrlI4ЍnqT moDUV1 hGeZU2YՏְF9!ftJL:M aFgӎ $]%YTnӺ4I>BCKcfa3 xjFYMF/a=fZKKK%0Dr3ˤUbS&;?Ο{W=V:as g"c0_"0?4eX8ai 'Ea&Yc_u8 G vH 4QZ j-ʽGrQwգrUKvcSjE*g8X2]f>d9\ Z,Tb_%w/o~UcV 2΄0X4JJZ6Oh5j>}L/y3)cPL^-m vct!A5X\$F+JaU7tkt i2 "0'Z: bZ1/}kC. b S{Xʐx 3I n:$ l,WPtcV! '/B?5/掸2 _33yR=5_|G7-S%%a~\{1/[}v+B(qA;J^x<`0mW~Ģ!0zko9Ge屚i6FcWDNgm~.]pA.ȣk!pfizεvooνà9Wsɸhe2ZsR>!Ug^n!^nEz_yNIRvj,DI+s*bpHdbT?2ͽ]BE2[+g4q&r#<V Xİ4DŽ >52GbB00aTT K̠<if-,\Xk FͅCFi DC G*!ٷ,Is 癐.|ASQ*B0hq :}fLl$=blcc"/ n CֹlN,1Xqina/s2XW θq X٨s[ޙx{Ҵѣk<8bfEɘ#Sr6 k/y7xl0[]3gtB]v4p- j4S)?OW/tOI5y8OWS֙a@LhS}5qq "*J*SI U~ޯ/^*c|G<5c>Ɨ.A{~Z7&A>fG?S.d b!cJZXy>pb,2ҥ$'SMG*Pֈtc!] KH E N_HH Z4adғqjzIo Ny)O2tvMdHɐZ|PM~!ӀS'&2dHGP> arv Aŗ_|52 i7^wԟ|=q}䤫°k}úU5"䤫O2ە!'C8Jj$C$;S$Ca5!'C$JjƧh vc'>PvMd:|@;;a5(,Bts=8ܥTQM$5oq Yf/V~%mOb4oFw^DFzzD"InMX`$)w 66BSz󧆧5 1_#:sy&;MɃAN8L&20$u&Z[l.Ed"~3 !'+!^l{<51HTM!SxÍЧ[D/qu\5N}U鳋xG RYS1u( '[vM,_=>yP.0 }rZ$n*qHwqtiG k˵c8SS.ۖT6:mMz| { LaťwdQ#|2`cWMX#Fc@o1/i*|1Xq& $(Ew/WP_(Y 7_Rk flyJ<wgPZ/azWP=x`鋆Ss$HT^Td^۶Cް~2H0qKFt6ϑAZa}L[2BFd6(Xqd3?Yxs1؟PԜ]^R Uf +-_TFౠ[$Ӣ ÷iLȸDưn!.10)5Ǻ4%^MʉWjVywkR*"x _IZWԆf^C< :XIq;zZ+c85moȵHIi Y-m)y@Q8V 8{(F jHqĖw#ifDn:wSzz9#DnN=Jw0Ed UB1-^~#c\V !ĵ߾[j ~TeZj뜰r޺3x5X"f.2!Ew#R2s ?<)6dU"%p[\*ihq6*-dSR WŢK'˭BU,vIl29&zOzOf಩v| t-U<&Apyװ\ff>\xYy /;`)f8A*YUw3 &rU`]z0[/nݸthyYD_ J%KfCyss_Fϗmv^49./>{oG zhh!(۴k\pO7Nw,qw mofx >^̖%J6x8/i>O_ fj ^{4+?b_w W7%=fKwTܜhe>%p\Ea: ֫^/y9oW:>jˌW=s;' \4& ѼH͇]јO̿;@fL\,I} fl*2^CA|ۣ!#POTT3D/L 5NwAwޛ~|7'l ywT~sVG9,i7,ߜmF\\` ]9$у8yJMAb؊7)O!Dw󎄥ig?vKuUfK`}I>A?d;gCQ9|P-aA":0&<spmn`<պn]s3jpsbNvcy3*5\tWdҶYyccxEe*-楦,yƻJM-_}_?\dJM#V|7ʟO ƕCmWrkC[IfI8ۼ-vEQE-[](w Z-$hJ)A*2Nf uG.|s&Zqqs}+`H[&2m3cRAv5~| n@8:wT 7zp%/z1[l2ȋrYYf?> 8dY"Int*+RxBBQ y[jQQtl6Vݰ=ZZ \aU]୯u$M0?5:ȷ"&{˟0\PPk:cNd:"??&>d^Qw驻^p;9%\㉎;Y"QwD>> J}DA'b 4{:r#D>8d5RvZhHkv铕P%J>BdlPFcQ'~r@r>FBoWOB|>fk.i};>'D'`OY(f1uЭ{J@'ƖL6:A#t: wEA쓙y2G}9żs[\[ݳ뛋]woeMJeo`-:Fu?wm#clh} ^vL23XFlȒFG_5H $_uUmw[{ 34LP:+Cc!k+{QWr~KWvF0gd~4Duiw`T|g&L߀~߳@5+ ۻl*,ނX+E&~zp3~LS9F{:,J̸u~˂*k,ue3O`>ÆByC3v2H* 1"(!J%U ]gA50K{ FXHtkx~ arB ;g#/U޶%dV%[_%%S9I[_%[_/9f_S×@:L;E$ L1'S%9֗%3Y> Ć E OM^ՎG-XXH'c?l~(*Ūb#bǛmc1Vc팱(O{#R"T n7-e&iBR* ؟!*cd dPC\֡a,\j1Nn8:\~̹J")#ݸqj{+p:pJq`Oř͓8N77'l! %T+f( tp['y\&sdSw L7Z0 (ʌUSNaBFX@hŵRG_i 7SdQ5Wa1 Ƅ jKHȸIU"$Kn}Qv >wߚ vKLZä;z xY,|HOT9bS}O R iPYhQ_GY/GSz *\Xp ?iR%ZE*h'!džd3HP(E G$|BǸf{;Gݍ;NdN2+/>:62¹ġ- C-(PԦhb̟a,mvOE^e?zz=_?[ÅT*Q2d'!qJ%15,BJQaku!$g?K`*l"bF"ZdE^qE^dE[:\dHUHiBf57ҚsfEJl}"3)k{Y+.`Ş0:ǔM*).ALNWJ}}(&ʐ()cM#e8Iৌ"#+J(lS%8*LԔ#cf)M51 qTņ iMjw )!D&NUS+1Qbch}L}J(nģ886rLb SXAh `WkbcZaB4 FFl 8M8SI0tLJ2y1ֵrB_7?Ƶv:71ӭs͋1 rKߵV::'mF%EURT<%]5\}չqʕ6$5!c ^*G BXZ|rOƪv|)l"F4Bl"9f$ȶKВI BRڨM3c%٩WfNcϦ0N&RS(8b= ]FF:oMY.A`֛$e}*1䴦~qHĒEa?F ѳihQy7nYkj ,,K_ٻZK׋? \q\<=- b@+<\ZcB%0Kpi`&G$ wSSwO[1y۳FXrp@.?-<+-?pg>;B4˿ ΪVak/0C09yf`1:;'ؿ ݢcLO^ؐps|]+rW5D&uMYd5ō1u?oW%;qrX8gXxK9A JyY%嘠BW#%k:&GK-lw$QǸK_쨛f PdJ}$e4N5r  F9qa qB0<9['n|1y8$I^mI'X!_qN.~-g\IoV/z,{wbzSZ8Zo;A۸m) cpg[Z'K}8J m @AL#"4G'oP%ƴ;x5"92]M􃾫e+ud)׬:ةl Lx=FYBEg.Qm(VI) X NV.:/00Mف+NP ͰL" ݝmn0s?+8<3*7V[֝-.ipC  !7 o vۼx 0녧#,LS,Okl1L8TZ|x\K3|+Q>|x҄F#i9hw%Iu)غwvp{hݺ$uX966fX Hׄ{^{ QO* R`\L؉)$*nE]7MYWfjʥ;yN%‚Z|hnu˰!j5HWEVԫMìNڛ㯢';o8VeG[\TzWU0jEj+ZԫdITPPMW'zK1f)2%+jWyu1&ӜUd:ۈZقQlWOS0^|.2\.p?+{|䎠y2-XDфa1M|$y)`,&QA*M/kL.=7vR2VELBRڭԂ"U}Ѡѕ?]/GHhGȽU*iipXƨT偏)2hӃy+.Oo 2/F` 1Cj!}m3}3fq+I٦{#|ExF Bx9߆vېnmC^!g76D[$FF 5F),L$!)r}:,D c.H_52L:Y8';>ClMϦ֮OnRkA<ś3FPP<;: AfcWwߧvIpHN7 LST,ՙC8(}'}%/z8`0fمA-ݔah&g׍.fp+b*Kj&o!}PxAWju]'ç0T[,m" 1DS[m1ObCHX'FӄS*,X,WŷolrFVn|eYK:Q#ĉJIUTIJDFD(ב̊8dD0JđdJGk 'VEeZ?DG)(!FXncke"6ʰ@n? 3 Bg!g6{\0>H rİP&m}>Zb 7ж܍Wh K:Z17)tHܒZCAz1\I!Yri=eK =z5B'! cգkn}B¯KY5fP!v0.jgbϺk̯UKZUJ#?T_dAG}2qLt-"%`C>շhqquStz7 4cJJr re4qqEa~v+^ w8Ђ, & xbw]ÿsL wmm$~ dWdqI^60j +K2Ez俧q8C=6ٜn]]kzu2: 5SSj9/jl^~b/kL1ȏ5THe<0pN-'+ Úԓ7wG=6BGt@@ɻH\]}ց%/+ks6o-/xA(ճf@TFwn-^ 9Vx5# 8FC.~ʶ#N)9VN'?/>,Ҵ]e\5A3k`Y2/pwV>}}X]Sk]ZcSc=hcWy;Yuj"=&޳:*cmN5qls\5VNrJ*mThZ*@G[x*Y,ςn%qR#P|Y臢9elh=OE4^!QBY+/E-. =e\ox%{A3)(Os4sIA)ojXa$|H%h7?6z۽sOtdt6å7nN ˬf΁I[S^㼑,93͉*ō>^zvZ"7T8*V%%!Z?vl@T N|&Szj.O3t,W a0 I0!Zhox> @>YsƱRt۴KFp 3<@aKo[)E'P! 9X "f3N]c;j!R_2>0*U/> mC hAoܵ9bEo_Be#7BD_D/ ~X3!%ɝZ:VN(Unz[L.iSo729F\P:XNV&RKW ExD#YK]dV.Dbk- p^-1%j,F EJGMcU +n fS1<ӵ8}$4ƌ"ʦa۔sH:uib %'r"e**X&s(L,AQ kKk*g~}ܔFE=f6E?ݩ/h(Ro|(΃Bch[[(b gerzRd(P&F:qfcҢ셡A'{HXJ˹CV *7'D>ilE#wVD+R` H(hB$g1Y_ G[b+}  .l>"Cթ2`ע|]z6{ &W!ݯo.WH_rqiLOY.7Uu*w|dYk R5_e/o[Y8J,UzTR9A %ZAOJ4hq!35JNǒ4MojtO%p+w]2>-q}C wYGPʁMUff~'.CYƣHP`ަa.V+mZn6_paoW4{[ o7sּe/>M/57[ͨ3CPf"M ijիZ} ˈ[\lmA9A.aBb0(:83uA-c~&Phu%#ZW4pά~ ~r;3ȵS 418c#neO:1Y X)HU{ިuԛ/m1jbh]rvN}o(NSZկWY?sr ]{c [n3)i!q(=dVD¹D"5U69K}eXҭ{PԺvOOa0kV;T 1$#`]mU{ jmrk@IJ >iOBDGX!0u>R*"tWs>k'z*уg%%ŜU<X%3$9A M>2NvhγzۢCH;`Kܛ>|%5/ȱ=5LKyD!6A"DR2h %MAx 0J v@4X sհhqtc/Lq8r9(bPR CCAuذUtXzx?zjQ503k҄{)S/xdei{ܔF2nvTYKIO09>_k'](r=']а픆|L*Dq EqzB=dNtkld(;Za@NkU:oZuǹQͧ1gϗVq:}f ۋ˼kUnâU6s~AN;hSŮ'c݆*?JFCwd߇b'Pg E}@Aߦvh\=ghVc> Q^.8_9zdN˯WEw_[}'"F@M'OA-B.Gi|Ϩ LQjFN*n ::6--=,$OY>cy=l'sRB)|i31 %SCN(bDy;\@ .: 5l Q>,ѢsngKh.q0z9^vF@rj^Zl'`;5ae^hmQE|YQ#W6IřbCLDd|S!ùZn{&5Q{pnZQJwvQJ~Zʆ2@FCm/Zp"ۏ?gQ*Ѥr8k=aYj9pmZF%!.] ʚ/ |ݼ?付6k7@7' c`{3VLvnZᛲ;M] H$4pf$t6qzf VkmmQ9j|"5 beCQ<|GR^MPN~e`( $4K0)(:_L d~cO"?>砮~/Ћ,9UQɵCHrnr- A11@VH5 L.ІhWd:dӰr6la3֩Ix#2lxbD0FE6w0e5-Pv⇙Vd{N]s5$ݮ̉7uMF fy9 GeN ٰh|yq!m %86" JFy8Lyn*ZKqCA\eeߣL#[id0QB sP3#%2(RYtΑA1Q@UFamkU"Lj%G _sg7g!S\m9/kOJZ0(IAx5W)vޞzA9N2G)VJDpx<:DZޒsQRK%DCZ;Ĝ \0c\r(H%]i{ [3\̽pK5[م]cg?kpɝw*=4?.'1ƨYa1zVfvUDVػU2}#fTئ&dnMdRIzTnDzG$<@Q&#{15#v0sܼFWX?ݻ(WO*z 62#t`Oby\(NDAvw=aƐd iyR~=z3῰o_NvJpK gsz7GsB_ګCLfNeg (% R0$~q2A 8%D!Gdn~O!+,A380Pa'-Q@2;vF!~9`CNsY{'̚w@'o?%̗{nOҰ8c(2 Ȉ28fBJJy8 /TzYe;-jr协dj!#@&9';C{< pn]D)ԖfD~g[DJ{Xq@{1^ڔtpI̛Vm5FT*epFR%!|::/@4]C"\as-L9!#i`p70C^(F3`JY*buRǨR6R)Sw5)emH $*u8J}d0,hRy#:cԱnz_# {nɄZ6$䅋hLI8zغ1F% jaL6.RR[*Ѿ[ E@XLs,$(yA&RSB@)2@*2RZahX --KmuO1jK2(f ~ئz&!>2fpuЂAu@ ln^,O.h+2Q $mg< 2k HO}xsz[7?*~ (!l?uQڊ8Շ9rQ͔˚5+Q3',뱠Բ> 0: <^%O&$ xhD0-DsjIθ#sF2!󿑹FV;? qFjyXҨZ ЧZ2_Im L#&2ʤ!ƜQ@N+*X (U*: sA@Kȑ(D9i5cz(N8B2h!Qy?y.h(F;…R& 'cq̗bI:bsIߙ083$yn4:" $hr*͉d"(1I5? ܶaֈQUOA%UuQ{H-5]GmCB^&XkYr!cy`8F0=D PM# kAB^&TTX=#Eo*bD'u:m]dBc[EL!;Zb B$/h(tPۥǏ걙*0#d-i8jqKc儔*tPOVC6Q QW-K ;j._{GzmE#Jĥq_bDQ(vQT?I}U͠z p(nPLUt?hف.\z./th2H2Ͼ!ѽkyPw]`{k[cYOb!C쟫ҫKRY|kow(fï.ӫbAcHd*HZ 2IDEH([%-BmBӜIg^3Hz&Xpm p X iTQUH8Ig!AFKv?)a p0G%XJVpہS<ka`YS]UZ!#OZFSR 5Co1Y_Ne#~qR:+/I0JT|?!% o"f0Fbcq8@h4R5GM߻GJK1hjx1^pu)A2Sxm>vvuBmp#snl>{ L[56I}5K|Kwur)$]h1$` antOΏ"G:\SL)zhXsfg5/sbIuEQ@1Uy1(! s[wM MnfGk(vj6,PS]`[MH)Ѐ?e%B3g1FrEA7z\bHМHX HHR .OG1 ״iw;|]͙kyOvhbI: zAKMe|g/n6ʿkbT&S áS7fz=Lr`҄)43yT+$>=}^!1)夆Z !)4) 6`3Ƴŏ~qvšUƝC=[@]pE7BqPoZwkg )ý:G;ݽ}zg "v擶aq2_$Lm>WWWag q8Eh5sG@+gyg (%m?0c./]M]0~zxJ+峥PPid<WN+j˟%br~L){+ʯ۩Zq *?X_N)Q<c.#T4&0Ssz`&w1SS;e+URBB{Jċ.Aw->Lw?tB v  !:m˪A$逩AȄTmvPSw錢uA|V~_T"kU1zQowTXM apf I.|>ΉL5&1`"4Msž=Ff:?ariRi`+2DN`m6$*ԥ !ؖ '2\P d}$"RH.eS!W@-U>Xh& r3em#*8϶dl8{$ ӮЊi,`|Znr r$eD՘ A!ԀI-U'5 ~ɔ~uUl6ɞ1Ye9GGxsUP;PNZTU"sˮ˝/'[YޙGѯkO~6pܦ&T ]EZQG~+bPU>0͇^[e.Áv,qy"sM1_LRw lqXc^b/j$gϣ%pŰa`EDs/ \G\ђg Sϑ: SX y"%SuE/g-%.V!3ks=jib$/݀?V%\⥏˃;F%G&;[gK|E4B4SWS"T7\0,T7)BXxthOx\8Țu3|c \35$o&JETag=,X<͗4C]ݸe/>ewò`퍻]I60%lfK[s4?0ϾqJh'^^)\tG$t rvS~\9rz.>,r;~4kIJtdl5rgx*bJZ~_d\US'sd_?QU8ϥ͙9aUU7P8N?^-P .ѷ)akYLeڀXEĮ՟/VwMnƼ2uE[EmZ9E(Y\7_.bbY'چ'\2sfM4cLR\"A2')|jT"d5uc vW yiLXŧpC͔ӄ9MI#Lw;O;Pa,RH8Rɂt > BVu:*Ÿjڟ8wВ@SwJG R*ݏ;y G=y(0('Y8MCt.dxLoX܃ \-C0.K)I6KOiKa.1+hBڶU.8HLIԀdE%nY亀6.$hvȑlL|d(n\P}[ .Y9j f5| b1n!`]Y>U︰lU︠{q􎫀䂫7˧;Gq4j(CD,[Uck/SUk )l=7xs9 *z/њ@ T)8j4V)yקw^ ~?6_¦N\hEA5u0HLahsFeܤbM5zv>KRD\;H ئ& wHH/*KLSfMFtDgfL[9ỵ3VX S 0/ ÖbYbJ\13g-o]`B!(wJb&LQ("LKK1͙MEM(+x̄3e6,%92; f gF* :0X["lqdžp8.oR ʧO rpMAwd,+bf`B?s&)#0Y)y`D efq T{ܬq)U!Cݔm/Ҹ|続8MaE3e穦RgQfUĜ Cre!EF \)]o8i <D꾥?)|Ltv_UHۙ0%RaexPQC{Njדi&Hb༉g'A` j %֩^B[ׂvv`RBGHJ )Q)DH1,xH8%!zf#iK8ECpa짓xAz!@&`'9: 1zLVt,"M]!W),Adl"ĊnT E`R)f#ky_ qb?Y8h(ukC "Doɉ Q< bj=F IRA`4 sYIhܴ#i|l J!~A%baYSO1]6X-AGPsCP%kQ^ԫV%BTGHs (N<0_Lɂ1T65.d7{1ߜ~hF`DU֫揿J"J]zVZpњDg|c<{3߿1O/ɼ]90 c~poNGO~aӐ;d殿}/aϫ@ ȧsioipǢ׍7~ 6|)F e.>5*ߠHX/p$Aߍ(ӏ5;`:R"@it,(f شii%1!$ rLIfT?F]r:~dvoF9/ٷϪ0T|݂\S`sNO<^q Z/ٴfuf$ 2'c S{ڥcLZ:șF`jq* !E9?RP/}FYFjD#7ώPJi긤VY1B{a tX_vQʨԌL@^jL9?BQäy\C}JvȺ0\+_ xN1hT\F5ш^"$b1R/Ua=B.Ta_J/p֨ \46y2. m82|(&`o<Q2筦RK7㨈eJ+#O?x2_7/kۛ xdRvgO"'1 9}F8^^J\(:vI%S[(O$Icc; ];},ݴ>_\3nOw[<ͱ=iJշ6KRw۳G86*)6< yP&Y7ŀP/HjoĿ_>Z$3d6 ^n_*伅YvdZ?3zȮIׁ@u $ L/TaXRTW{Kly0pts-̍WjlM77~4ִ[;B^!Q1yMzgjMgZZ*v*9Q]ATKJ T 9j)Q]rh E0ڠpjhڊAkt1Ád"܀Xs`\Z|U(aUQyy-GL&>|k&ԛy荂ì,\ٍLf~8WTUaRZ(͇a(;R'Q#:E4G"%s87Id`Qɀ ĖXFS%SP_aSK"Jk]֐rI $$\E91QZP'D57Q(f!/"rF2y &(AkD|4pt`-EBh0 \cX_RB5v4zK lԧsl4Ӊ{EDK Ƒ6J%IJB9|ĆoUnJy7W'uIP*=:+4iJ1&Y?3R[ߓuAW-eS_r_&F.U6B4/h}xӜ/e!r!%O=.o{-A|.7I˓|o.aMXQk1f ?tvx`I'bv⿼(?Qy3Uҟ\{С&bǥie՞(9d|)tuz)m 8SCSֳxko8 kԀIb%f@`ko#񋨽”Bl(OTԔh߲%g1]DxA#;rR2$~Td귥ˤdBR.H87l}|hg@tH'37+I@VE )+'n\0ƙRb*pJ1nߝb(ۻ45m5-%+=:aL1Q/2 ;G6 l^ZNEa1ڟa\<f.rIyhlrPl2 @\+]2-$iE/xqdE5,@jKQ6"k9. ʓϵn|vd*A-e(E]ia)kN/}Y*/,k i |>-+`_r$unp,5'inc>tyǖuseOTV͵03I4PTl^~?x!;dw}sI w"J&/5ϓeOe&Z#p6gC ` DgDn͊,m&rdU-{BgmrJ#?n7 _6AXd\3$%Q%q8n(V$M̙> Skj , {jׇsA[ߎb[\8qEڟN&K k{E,ŦeM(5C?/s%!]8ڐ1lNT4aI埥m gO9rUbOdxWX=^3.!uЩ:J֬/ۥ?Na>:PqJ(nlo? Q؁C-(C8j3؉y\FὟVzT,-ǔ WzC]~N ggK AKc3UJ'9C~J!W֬Cޕجs\;Hlқ^ܳh]J^`Lef[).0jHٴn|FժfS$ƓP&Dq]e5(Rق7SNK{vj X~RP]Ĉ2wh+˦9NG31 nXeUbFS ].>5@08'nC.d0zWcub24T]NUZJ0s,@IncQ ,bT'h.T0Ei@@6_^9s${8_c@Q;q pq%Ɠ~f_y~Qb`aS픭8 ~k$r<$>!KHpBW JA5 Uv |%]{Wv3 mBΩ9Bvs n=QM]KGZH!0]!gi#=/ىQ60G]T3ܨ:kRĘp JKo_bF ,8J([%fKKPX< WwYO5Y@:b'40qr.}}ta~x W_{y+mߢۛ^${[&Eڌ l{RR3Z~4yA_W6u4em$V;[bf`<_e'1CwϨJ B}Lc΁/+ =/sӛQs-Q?`GX,7kw Rлs sT@uPD)=PWd^Dɸ(W~J`@7x2(k"VR0N B.0eӻD'q9cG`)yG;YH,#?xNDr؟&Iszpڱ8C pFt_ULB6EMʢCŌ$!bKN0scӆq\8Qvnƅ5рޥM`6bǗG!HgVC%fv|+ x9av K4^"ddt2`W:@s\Mc9L'qsg;oR ۗzh&2dN X+]Uu2YZDB8yjTxY2BNNΜ]LP'R@NJ} jNxr}VgTꨮ;CD9 ܔG~%wZrVqO͑BΓ6=&́Wg0&X (/J x9"Bm;},$.8I4T;,N0S^Nbo w5&z|>-ϒ2vPd"xM϶*)Ij#{:PNN-#/[ަs۳q|Y 0}B uܑIH u9^=G3Kwbnyc(15Vb2D:B"/aO…2L~觾 Y=CFEMJ,RfA]W4$%vi6^637dr,M&/Pکww/^p^Ue`= \vwjh_Q4ҏCmM}Q՘u4T zSjX%J̹{z\ A"Ǽ)AT! '8'j'ꄌhy_Z~f{J;6|bt\wu[RWpx ktNl9% P86;f}Y7nuFD> y5WH.5lwQrJ"A]?L) AK:tD2Sc6:ݼza8~OӛW3d -i{7^*z3yRn(B_!!| P彯[Dt_.Xꭄ`ʼB #HX*H08db? @ YH3|is3&e a f҅K Om&U~q\G`V6TLC jpU5 w ק9J Gxkdx%'.F4FHm\GC u܌$jd.r+ O0I9gv!3qk=(Ev6UE9F;|6[zVPpQ`JTu9lCQSթǪ :Ub g#&અ":_*-w5XZ !$uq\1 `r#3-7)A߿'IpfK6; }%`90vc9ե]]kI:iCՊB Um}P?Gly)`tdеP7P@貂b @9nt=~rP; "(s-8r._-P"jk8̭oJ$wxi\P>ґ*}Nm0@'O#eӝ;*w>\b%Is HFb_BRb | &Bi rvA]~"uBRF Т(t[C8oXRzL(9ReNMMM+[' ,PhCX\7f 8xc.eEAmCIHЀ:6RlJ :W;le>M-JU@]~I9]օG0E@7GF)t81I*yItt쳁F`5|RׇSs$!"C6 c4փjm i]e(Z@)9K̎.hn;LxsR%'՜Lv>j)~:3zW*W<7XO{W啺xk'`ߝ4;W`pGW`k5A۽J'S6^6K8Ο3v{C}ʖx#EtO&|/yPv+!$wb8(T*s sbtjH+[~KoVN_z>ǓU >ި[V5]׫k!#%'nzϊb3ZDU[Hφn7FR[]R*l_%qw %R`锤IT-u-974Gq]xoz"y3d%d;bRΒ?i9/vDtA{9E?{oAI?]'X3ŅZp#P/״:7t7_ZG|-θk/\r_7(wOtƟ'/!FHH*!Rm_ǞW*wn"mTRXap3285]\"wʳ!\vJ[30_޷- ;_^tK֫cU,u!vvtsmuӾ;LҿA6IJ)#Uƈ9buW2+ _ԟK}eMil(OJxaE\arYۏ?E'dK ZJo  a }'(@}IFLPFs";FRt#. |/{|)sSR-#2t_1x\t_K|   )i- RJRϺ2B]hPx'`fCe%w(qJbP(k ?ۭ#wjsvX'ҭ{^no~JsewJO>zzsBZ0 D8k L# Q7^s,y|.# bJwjE%y_#J l`U≒E_޴D *E^R Ud5,fidZ 9/?ԿuT%c=Oҷ0[d'^fA1n-vQ?I2V&_=/nV6̾HV[$נ.^έ$y-c[-k<FI~4yw`Xڐ_ifA|h]FF~y8WF6sx*RAmmSP[!Ag}J a]"J~콻Q?wzwXٚ[V&kg/kOd ~Ү 9kPM"$pR]H}dgf Z-TߢU&J|ikXݹcUH¡! G! I$iJ0BQa}# eznNG`B)xI^vZr|V,Y[&ë⳹pV/ٳAxf- @N/-;ʅc19r| ^^7ܞ1 a/ieщc8Yz]^0Y,kzՏ<ټ' Q" E\mqBr Q;[ps&{ R3q=$:r+[ a()crz.FSAĆ@0` ~(X(BaN: ]CXwqr.}cga-|WO{*60~~5Nf}#1d$^ve~Sm!X~9Bw>2I |\eSF(7;fJEcyS?IH R5=rZc*V(T<|=ip9m+߉Tf9聼 ЊkF=X﹭{ؐ,?8a5gX0W,Faaa+Ջq\ꑫ΋ $꧋Lzƭj0NnR#-׀`lMb[;a"@pm2"G(Ljngo?d${5v6ȡR˫Wv7V~~n@%I06&dB%'/]oW~зU p%,iu=řh/K vQObHe'TDAvIQ(Fjlp@cJFS:9=g:fE"\S%o*>[=TjWN_F~F~1 $n'boI3Ƨ%n;&+G 8(9)8RD^'9rՉAD)+;sF6f߳7BoNmޔ[.vc95.:ũRmxw93k<|o?koSrǾ;FtpNn3],޾D"SR\-.4W' w{rXƤ݇4-'&V~Vb]oeٽǔ_8>{ TS;25ĊI-*iHl~7kZ2-@3oP͖.t.e)a}ݸۨvJSQBe<{;R ͻ++qE4u5ԪVbW3Prqй3 1HB- Q`}Y9 i+8U2B*(G"@Y_?ydNG2T6*|IGC o{y%t<=u!Uø5<x|-w)mꗩ+t9?6^#ͦ82]^{V_6+Jy Db֨JªAA!+:8_.S>]-D|SݟOK^ 5#ȪUuZ88ۭ~i%0B=ߵ>(j) dbC+LGUȑ(Aq~P:9U9o})2Tpܩ>R})1Jz^4ӛr=+κ*#ܷ2˟~U$CuA!ZྷszR^=Jo z)QFLh4Խ!RaRzyr ֐F4_{X 0()s>ppn$I&=Vc5hIjjŖm"^)~>|C|W\-^ w&N3j|Đ1 1d#>rwj0t=&i-J{Y4%+x"h X7N}DX &0< yzښ4r -mùX&,]oɩ)z K<1 ORN581bwWO\1\bĕ}#ӨTkO8`iA:3᧰@?fޣ6 b\bw%")tJ2W+$j,Ƞ_E 6]Z5ku!r@n-Iꃈ_[Dd d+V֒-Ȋ6ile%U.ԩ9RĂ4r_W;oAץ'df\n[l3'@hfs 9"+$4 aǀת6+iSFb1XFfE#DQHب]T:0E-'D3( X,yFQwa Ԯ{k&>\_.~A|ϾF^b&nÚIc̈́#HENpP ]T_u2 A:$v8"N d}ba$%dzcX]u5֧"%/>~*^.xJd\N DWkJ?©:;:Aѻ1k`rt PռVlKaGWب))} q6MݑXNLu;h4v S{ҰEؽjiO#wc4#LLБ%4[v >%_\=iy:R~tZ T,Lu# 8 Ci`~+"|3!lm>mΈ@< <2#|p08(#h?NZ.|xܵl)2A~Y{ΦN Zut2(x CBI=ш9WN-^XEPڦ^ݬ)$@; c# ÷ V!mJ0$hD$gZpmj R!G/a` ǩ+_F?}8L0e/ah:nX9sҪ.a83&b̨CܢYrDJ)=8;]gU89+幃8tc'/sC+Sl)E67,3aXQFTŠAd!4!-BEvJ.&WJC  kX/XE)8`K48@fd18P͇twί xW;o-`nhcƐkq8 h$_{~bM{&\zp=0Z[mʳV£r(c5JVkR6\9 fJF!8"F[5`34ک-cJVL_ԌG艚FeFWOM1F9 sM9ߖ.,^{N%ܪvɺbZ;{swP%γYiýÕ_?^§ϻ( m]O3ݙ_~P3qr/$e%:-Aʹܬ\Qj,䕛hMu,U ,[d+ qσ #Ai,GJ7EgDyFtt%GQ%$F( +gh*@SSa, {-P91c,?O4ݤU-9:YZhR{OV-{PdeYDV>_!M,.PSzϭ*`D=rS&EXRj5Q7[; 4/'Mn҆ޝbVȔn:xX\ 7ݦ!.7`PHH4BbN#NXVdDZ@|) G+)5t'eQ5eZm%Yy rs(|mR4;G! 5+"* ksI)GI,!AP :XȨw3ŀhF$hͿnusJ}-B/MH=p[qw?ɾǽzztc{ܔbZ+4F5zzѼ2-kf=ѕSh/OsN qݏLbL08Ldf0m8 7.Ƴϋ=ߑZv+'/ '6Y:'4B^ؔFhI;MԄޭ)mwQiohV\ևr-z\ӡwnNimۄ)SAGnn}X+7"ZI Uw%gcHdL!(I"iRာq0jf Ti:^Ib%a1uj2:G fX=6YB+(HBη3K1q.mSJh_y y&ZbS$MY̖b:*,:t-[pƳD҃u` N{MN8 Q Q iG)L{AAdK]I^ d8p]OA[)'YTVL»a1Bh®[ .af<}F+؞?Yc1B)9˶6,LQܝ>,䕛hM)55w6w+ tJh&w֪݊;WnE6W MY bUV[J$kWZ}=X+76)lT^MU l|T}*~[T 1!%\W87\LK1+PޕaÆAzp5%%1~DX!qXt‚Qlhחyg_cIH+wyl |L2\xU$V~͕E` 'Wcm`S6}rYI ,+>eV(T]k/|Z]%%}VY)\6Y?Վo:xl-L2|l ^2pYgV)>7 J'OR16'/\ͩ|d{؜_* X-`l9L&fgچĚAyk7W;Id8pOnoM%.}gTߓ7~t@fag{^IpC\b^>?±E|vǭM?; G5h7bIa-E(f6-!eD>b6TmQj~B킑 pHjP8`~q>t~aqa2/:jyɄ, 5~NJ3t<:h6jl['f0>0qD_&סoFcm'8i:e0@ boN;=7|oa& +Jzg%Io@\DZ&]hB(@L &NZ1Ի K(UayƑRбsof+Ԁi|3[(=>W>ژ7!!ϒ6] $|w.pu^HrW BTq4+AB#탮^ک5_3A}EL4:-ԖiS2jL8o@JF*1ǺI?0 t_|Ŝ-s<{FqV@vvs;Oa@}ί ++6g5j'Y&=q+.i2*U&j"8V]D"@H|@;XiM`דp12ӹG;Ԅ[m]:AQN | q1 MIIJ=Jբ!ypvP?a՘9;$*&g lW[3$ hF ΪNUAEHUv$A#;ά|d mFƋ v%W8%q4Uʔe ISmW"]zo%] bG"%ϻ=93g$tgOը׉aі9^0Q{XN__.L4HQ7T; Z}G mN==zw 0]0IW;Q'9jCVE\բ۫EkykE[O `'Lj# eFig2#(5<ȏDy̕'Q#rގ L>_3NXG0F&ɛ`$o)) 1Np-!)c^5rY+.#+sLDh:QSM'[A弍6L*nSz;r UBσ#^/85BG ~5MB ЄE $H09b CIetș BI@m+Ԉ 6c)vzyp+,<&d2L3v #$Hx/UfZ`3ql2/#%a0 ( <8noyGB (\,u9rWMk^X >!zcMFBlՃAid =HЃ!DŽZъd ](zPv!"V(92%- >rͥҏpB0%\ѶrLXM="%Ou)N6^r=:$ `_@\ںӑ4@CM:DA0W1)888 6  ,V9U@]JH2jGK?Aj;V@\%#/ %NJaDИGw Wʠ#*/޵,$FF0Њ$ Z"mŘ 'ny!+.6Mhg^:m܄Sv%jYA>9ӟ}96 +ZAaQ 2'|W -"!@: yd5FK%WP-[(t{~O|q2qs0WK~\( M/i ~rYz l:M30?SPA &aʞuI#q85$xRcx6 #.pXcxA$f)D룙h "X( UE"kAk> >B(AǀapA,;ьw GL}g`neZ@zEy$κUdz֥Ks{k*̵ҪV+tƐ4I.Hb3ժoVގ{*igey 9UY65{3̙ K]􊢽 OUk]M:Wa[Ʈ &-vǮ+gki("J5&{35RwL8ʥc57Jma+}:·f{ft#09'jb({ )`$<`ԟ5n"9Θ k|0gamNA˴L=*'-x}s&0G` ~x/tRIcq~bsd=4^$So<߻wtdw)Hfw}ig,3*Ն P&:o+q8Odz |ZӲ?R!u.,|=U'Ά A3QxhNz-8?Yuꧏj0NhBmmegţGe O1<D1Ԯ{,u[蹛w=Ic8HüH -օߚcݘ 'JT,4&IJ;l%Nu佰wJ !KLH C꠴qsjG\fXdJԞjrK MB{K9cM=b ZA t_%rOQKnƺ"j Wn]v;Ox *o*tRHkMߙpMl ǣR&!YR+WOf|e/-4b qr8j>ٜ:+%1:G](; '.oVBs7d<47}~V*d~2 [{recAbgIn fn:]^\`"×Jju1s K%O\. =Dfj6 rTT:_Y EZB"*IkVˌ25|M|ʤ)Z.2" L<yǹC/۱^܈9QA(EF˸F˗ "Rt a{={oq\@-08_NjJ ,[,ssOߒwm͍8evwI򲙞Jj2S)it%>%Gߑ{WhZ*<[2@9rrvV|b +Mʲ1gc%^iV"\PM wr_d jʣ8l/2 %IPد '6D-c_\``ȓbN>B̆GA(W1NjS  7Pcy.IzCf`QY ^&)j l=|`U+s.{ym]cO.Խwo.p=SbzځUJi4"ǩJQ~4\}{4&8+IhI,-Qc]6lT|WBy:wZ@;rU4eZD׾kp/π Laǭ*gL(ZvЃxFw>': 窉p6=[L]=Ejzyy7 ӛbu>.ˢFH;њMEݶ0q@O݈ex}Ob rL h7c9t\4  ︽+5[am&;kU&qdd[f+H4 CtH/hA)t/f6#}h)QCYI/p_esEtVZ#\t32 ,gR 04R [SVGDj@a#% ;;)RX߇nJLm ߾|!LSN{丁k A3ݳp%jcxq`mv|(0'[rF=KIH[>#,oAl'i8c-Дa%ޮ0 %4vp׷VoedM֝3;SHݭ}{x󳚷M9Ye0ܼFM>Ķ2(qiVUIp;^$F*m%# koSk4iۑ\ӝhg<D3xqdc o X4ps ArW2' m >ZR>hT* %}d\Ght>^&mM5רrX3V |'q]AP-n%B|" `2bo*y% ?@D!Vw񾣢g/cv}1E97^Z~jqΣZ\zxE?v. Jnk(.?&#ˌ=VC~խ7zb\O(ITY s7 =Ľ;ǰfpCT!&k==5iMHEﲒ=)&`@nV/] 8B&`lw?尺{W)G)C繧i^σ;R'7[͖+OnI@}Uz-@Uϡ2)U֓](V+.*&:ɣ{ߞ|}X{S:dhpy>^d bRN!Pu! Ymx]cּ9K#ػ~^Z9]NFu3yX4 vޱVZcOn5o. #K%O6aD+sM+2|y# lX~ooϹBk5B,ߺt;o7̓\3e.&D(+dH w,7jԔR|WLW5|\&_ʎ:)0tE!#6үk%iY˯UߵL^ꗻ%O_VJ=5x2MS0v.EDFr_61V+Jbv4Y$_;oRr~u_z\azra,tJ2ZE v 0h1deg Gٝ\EE-$*|:i6Zt3{ͧZ$6|kU(ywl2 vd,f*q\VJU⸿Jg0.9m,d2SJo1hŔ\ 0zfif|kVLmJ/'Uq\A;$1VXaNQZkE Ru=8洍`ѐIdfh9ӻ볌`W M%f hhcR%.I!bpNZƂJDTtLs!1TZc$1rk5d+l4!jI#Pq$LJmjڗړ}C7jLˎ,Dn [,2JPM9(O d@V\@)nU^|9K|:.'㛩+ג//yv&.0vɬГXVc_o{ ,_qr^_}莉)4:hRK'ȵ"̄J]r?_\PuÑ$ɊX- gK}j@C͢oǼv͘B€nLjw]jƯA;JFC4VXJHB*O)0Y(M,TV5ZIL Q>$hf5B.^7}̞$)1}" i!k]}BuB&r 8Xn]7\*7H]ڴ]|ݟ=_g ^Y>auj$3O_>P$%s嶢 ʐ",Ck{C dFֽ\2t jr˓2Z2s& #-O. bFGƴxو_7) O0*t-MA䐥Q!hJK!HyclwO /wn'zJ,'D#2&,0)8;ļd QQDヌo낿 hsv !dU9ssJ$,%sș\XnTf9)uP44:'ԑ@+r.zɞ=2_T1qSg/7⦞l&x-ŠLa"rU)SFhY?&yґ9=Amt$1L|Ņ sb&ZB!B]$ Xr@h@ } r%$C>Iϐ5='̤s!F\l "t^h0L Lcҕ-A|x]u><*7_0 (׀!ń^} \OSB8퉴Q9p4{ hq ROPkJtrSS}&KǯY[wt{/?|vS.PI!C9֓ef-Zb>Cx3_W&z{0HsLɣK͈"%Z'zf3Pi!9 O Mh@X(m!XW&yROMf~oB (ཎ:' $l@`H)o3z,prt]t~M{Q v=Pm{{qڗi^hTz^Cj44M\4yqcY]*.~s2#`G4;+o]zͦȒ[.Q3'wu6xh[+j.z_ẅZ#f$N f^j=>oͅ A:x5Bq#͵`ozC^kЯA0h.GM3|<SJ8taXCc k9Cgz_l/ē @ vlo=EVKݶd%uėd$Ub1 ~ފ:{ kƣf!PBQ,籌ST8Ewb_WiM'L~2=tXD.VVEMQHۍ#ֲQrDt(fFSP֢Qphѷ.d)ATlP6+UrG*xbVwB&M!C*%0b3=ihvHu@Ud+|&`ߢXs!Qh>+l1Hv|u7(?_2܎i@lwÖ{|>#VNעpr4{.-p}}kPm~vatnXrhmpd{6IbB<jd}tvjisޗ^ g2EjWy{a^ouӰdvoz,;M#3?h`U,lQ>e@B+ᵝgq1y{a5/YCul_ipO(|>U[Oe5OM,>-eիߓwwbUo:; bon4@M[w/0!^.1޼z/:͝ @qZ](zHG^za*8Z7m(z bt;_ ~ގ?>kdOFG Aޝ,m4N}PxǨz?'iZuQOO|YyA;L!( <[!X٣&b6׼j^>j%G_5svy,sjͷ2RC(ʻ۠M 2wCU#>KޞGpOМ c{y=ki׶c:='iᏣ'Y=8:LwukmvW`!{;}YN[ҩ-٢Xyl~='b|+*:El~vЃ>֚mρ<$-?K\I+l#/͸=O1t%:7T%}ͫ\ѳtq:w%:QQJG!~{459#`a^<y>mO*e-pY0ϑP2KmG7F`%.V8{ӧ{ӧ~ q _/ƽ qHeF1/XdBd' v$ʼz?](^WIi8ٛFK>hK?fC;xrhlp8. 8gȉǂJq̱>r [iY[O곁-9G^nPVzHFR5WZ=pTA>2QJ1[LTTRX]N:\vA{N7.@X~)@KF4iybayMn4?M&KWXz^sh sCM'Ȋd h*X~߁ Rsj.&7bʜe#+s?vB\bÆ֪Sͦ|Z#un. ~FA:RAquDŧZ,j3 R\:Zh*Z9GV^6|iIEjFjK3 E:֕DH4@^Q+8 Wb4z Fp] u^XQ+ֆ*Za} DxVTzmb:UM@4zYA`F L  \' tE8t,9tԥҹ#H8XBP 4@x/dÐצӹؐ C++>kR㛎~Ex$gk>JyR]yȗ`RiY ªbؓk(= ,Q}t:[HD:{~"*I Ou΁tijRa|&K@_[(- 8Y!-LJ`W5ͅT4B43̜fCHeueIרS)O :uy ; HsIf>:LkZuBW_ WyzW ^kT'B`Y=)s(8[i"er<ɑi6Xk:3WS1 H,g텁LÇl`ex:^䇿o|18Ɠ`j] /oB~5Yo6}ӐHWX 1zT] xs̸,TYO[SC /~w_~54cFﯯJOq驦VfFc")&B[Q 17@*=e>t B 坬 \wrc˧Os8wxmFfBJBD0Hς A!1ǂu:8S5+@jH k%1oI2&. ȶ%T9\B:b Nr-4bG*e[FZ4zJh3YחA7Ư5 Y*MW+RwZp^bQl L/"W]hGh%TK +a棴RY8"GܼWx}Xݎ5/4_+…B' C$\c¯ :/2jՙza,cI2UAS8a5Z2F.c;'M5=Ny]R t\yo 3#8^ xF$p'G& ^g 2SX[lLdr 2kc#a1Ւ к-"`>₷9O<ӀTAg5B0O(쨵peph [z('=/Z)'EqDp^K~{s,7%[~/0X 2 e|8=ޡ F+@X`RfCA ~AaHgc9v]Xxj~Yz e )_.AI52pn0ʰ5Ve*ek: t͏56( 3!%3Aʐ Jäw$r삵8v. wJ% K1*798ZaVhn@d!rb )R,W<4R P3A_Wbp547Q>}? 񟿵WICuzļh^+L񩽎6kN5D 6HN﵄^hS^+I Fɽ{|k)]|kw).ܗ rvA0_ 7.l'04[~g|X8fLg<R~a1xn&8ώ^Ʀ_ga*|/GBQ}^uA0f^0" HOvo̿ U#!g.Q2%d}Fx! EAщm֞s*qc-~vBB\DcdJaj7]Zn6nK圖HGJj*$E4B(:U5'('Lc1(#:mg91/ݪ3(*vӅZn6nc4g5ggnUHșhLIߎvc .hX ʈNlh9R6Jj*$E4F8bn_kQbPFtb8Fߑsʛ3kKj*$E4J h7vŠĶqvk/,Tƴ[||IV󹅻Mh3A-6*/ r\T!tL٨Br"%SRv>yGeb1(#:m+H5ϿnUHșhLa,VgzOy)ҙۙ[:@Jgֲo5xL|ٔ/[L\a^kDRneʭt&hFVAsEHyx)ʙZbrRV3AZuDI=[M 5U(傤\Jg;5fހ7PLЅ[RL P1s3!OEQ1 @Rs1W9]]g5m0g<Ř!,T_Y`S9Ř 5 Ls1W:Ǫw1fA4)Ɯb̕yb̀ Q)ƜbUJ])Ɯb̕IIb̂O9ĘAx2jb)\L`zmƘl:j1ES6cqb)\LHŘ9U<ŘSҙP=ݛ3Ws1W:1 *p1s3APJs1W:$Q1&OoYĘ%R=1.4ŘSҙ)_YZb)\LP'Wj1ƬHc>"/ƬISڙ D1kDY1K' 7ןf?˯6Ls.ލhv;/Q[%|?H3/o̭  z F 52Ve#.40 ek4@fI72jS4He-:d!S δt9Ja)twJPmV 5bpXB:(m|BКHXXZ@(JB(.8"L7}XT#$ܞ[ fwz?57Zm.#n P8GB3J=(px,] /or2ۦ/%2oܶQ%.$(BqQJBByk!TzElfck ڌË1B*as깵"___j575zO4l;h>"&DVh-Z't`a>;pf:ac_ ~ bmi`> fӰ#%G1%=$3Fcq`JnVcmf~5zQfg0wa=HqlGssl/9s˨NK,kZIgetg E`Otu~6ݪ6v,dz髏Sd 2me7RF+9eP.,F>|t=ksǑ/*YpbɹʎT|k$LP@,^ *"ݙ~MOwOwO :韃d%Jn_Bs *ثObŚ9^+'Y 9ӥ)RL=6V'PNQrE##C/8^'sG[ L'4Rl H&*Y՛W~wS4w}ԛF'gAR0`*?}$׭^=}rd }ViRhÑ!+’!KC*8,S>զQKt)^ ]a[juz0rS:S۟[k`>"0q8D=pjBTE,鉗ֱīEuL~vzU':괞]z{%`Q(yƗ!Kkx۹ϲ>,I IVY]K_Aⴗc1j}lV&ٶbˠVTu%U -\<g@B|uD&R^8)qp)]\z5z%4z =k30۲ q#(FX` ^)`/@I\`]Hpa )ԗKP6hu-R?bhߕ{wfF$~ߛ˳b"h d{YO}V,yn8 N&¸v+mDcb)1~:nt5 mTg  ('dž<*ľeҊDLf5(KV ^ R0^CZ1WҾ31 id`QUַzދ0!-x: ? _gGՒGRV4Ժ7aZejh?s>r܂rN0r=OWprSÁI_  |qg\| ;-+k2 h!F\4FJ HM1 M ӑDCIC݆Q~HdDSoUquZ+)NKZ*8ayŝt,{ 6&J984hrCC%mN9!KQx <\KQJgKQ8^XRi%kKYNAE+)pˈ"sR)Qj8| AH9*`]5`oHy1%{i\&tLc RnD'ٜs NAs d]":T'92ܕCj½P>w:Vlw jM˓o=h_d׵@}v>7 ͮaN3DdWhoWv,] w#"St]AI°".ha5vGgVcy~s|+k$Mjrw|trjda h)   bz*UӫV Ԛ[x6`KR(&z%*8ae*S7B*$;td0N7ʗb.Kȥb>)Jmvnv}|_  wt9 !3җP fᇷ <{7%vi|pK$~Çh|?F/+0?]/{=0MZ?uv*s@ uz-<`^x6܇6Ɯbhǩ(NJFlQbR0Υ "q6 %cB&Y7 VlcT![Y 'HV?7n-!&5A|)9 +[hXFtq8F᫘nM, Ip(fmWIL gNm.~UN }1I9 ~\uhl ލIiv3:D֝P! M[ȹJFWՂWr-,k .(&MgS,j-A㣆de:"Pb2| )t{4ZwtQKvRAҭu,(VKQ1zXPkXl`lY1oXrKolum\ܛ)n KuqnX؈mą1`N5~Zg̒Ţrw".m{{e9Z38Ш92"4nhS.E+G[ j65jnܾ@\zׅ Ki +Q._N|?p] d2w]e "= IHb @97@[S$eLBI0Ȩܻ x${\~q21+;VΔY;Vc'ð$nQ.9œ)``SK. Rށ17"<*2Fώ\4 t1?>b0G?6<138{;B5gY_Oq˕&ݝ3EۖGI`XG϶ iG7 ZFPv ' rv5s|EUpc,e&JCק0|Zq/BJk g!^a”(z54Uֿ~TVnze9\Λeg?&lVY[\]gNP@Aҝ/s}4cxa-E{0|ZU[TQ]X};ҔӶ(OiK`ЦqQ1j&+]8yO[sGp;'3|+0) x.Z`״!V_4x+mHЛK*CXa/^yJHLQ=Y")/1u&|6"22"22:m[u=n;l`ckKfΑ e[V{Rjʡӫ&$QP*KJsS oP4ם=uy݊Ҟn%UWϡ~(?afVal󼉸¢0ߏ:M9{{0׈5=[L=[JY|m-V O{3 '5tfrdN:w9$hhpam" 0Ke:yIѽeC% 蛾7ao}amz.|Rڅ^I bjSbp2t XC^Y$;\C̀o=ٛ\ kfK%B^ĆJ +hZ^W˒ŧx҈lۀYL}*Ktٗ=nIQ%Fp&ZbٙgČPvų %V ̾T)`\d?(Fvޙ/͟=nrH̙EYٶJsqSM{ŊRjDZ0>d 3,>w`<# D 3 '."L( e}Q/?\A:1SIł]s/M>sl&uYT^dUrLW7I||t*9er,C)kP PA(U1F),!ƹ%RhO`4Oy2''2L2mɷoB\I-[d;Ll훈V Rr|`Ap_67 k#F hјm}Q-k}i @TZ[a m ShcBrh, `Y$H=EF;(26ig t/_)<G{+!m<` .(JDTДE(P'J 6#P0a}TAc٢"٪YS,Ϻ cR U:ɺ+Zaaa< SOGe8Q p)+ѽEqV;Z<aJx#x4`P|C`fe е̙t|=˿7ӿGw'іA%Q 89b[)$#2_|=6YY`9E\n.8 l3G>'ߙiOӪ$pacA6Kh9їNqjǶRD_ɿf6$V fOIjIq&x]>w ϳ@ބcߪ}}T4VF.vqj== ң/L^߇i9@J˻`w_\:*QRI/I a+/׵@.]‹QOf%U碜6[w+doxƯ?\_DiABI3E_E K$ vdr/>yN81EY͋_zDDml !hl:ё#k;R͌L NцRIlHM7B۽ 8(v:j/ ;=fCTbdN8ǭf -8b?Qw7W!1}pX7!%Dz4JozJ+080e-3AVmI*D;GU[QaOLj٘F"PIAӑFR1AZ&v<|)¤}zcظ'1)A-TW%-uq^+-,f59rh)˶;`F8[ysȊ)0R٨s;MR^K ~o|-BoASq6|nj/7{#u%uɩl[Mh>s.n'-(Gxr{ˋ/k0fNhH/lIO/~ 'D&kվ>uN@:[R͉-ĝZe)T:,x4VgEKswp1hFu {9˕ppqQ9(\)$C:XӨGTȉ\m?Pso')d?;,3Ɵ+M/:TsJO4RSaR$t@'߭uԢ2!S-@Ddn ,*fʔ]!Ԃi.hGJ/&@O-"N橅 oQv 9h1fmZJJ^r7"*{j1dvtK*QRIJ渋Tmp( 4bb <)C>ڥRp`:)A5%QȪ0ݑKW:@~Jš˅K 2b+Et4RJ,(ϤO Nx B0IR<*+aL yN*AK)Rޮҽ8?y|0\_LBFrO x a- >kO2y=8iC &IKGCxXY>|2<}|(x6 $dزHQR} [XŽϒX1XZǓ/ePζpeɔbҧ3MkqNRΛ+{tmV!ʗu=H?0VX㧗og<1;t%}v>DSߏS޹ÙARz_Ʒ--LFK NNg77޺l:uy}0aQt}-P_+ 7֔r۰or+ voWB+^xǙmc}گ^ճGX*zP"_&O}|>.耍oL(E .c >hk#r IHB c a+B|SaW@a~8 IZv^.]-X.gcSNbFn^{^q-_OIϞITM=9P| Fy<$wB{2 8_w7\h-K47Ԛ帓r<ƒ QH|m`Q"S&>Msd4%7|_UȩA><Ç}gyV,< 5WNչ!0?iKNyl^wPϛWy(i m߅ɴ WhО\+*m/8e%W=W |jVJVo:e=nwqaڔyG:ڔYw֦\}[)_W4BҔ Cp'6GvME,qqv4s}Na[Nm9p Fr&yیkɭTM3R+},{2h\sFJEw $f7HhAB%*8c7H4b\zep$ QbPFaۅC5GVqٛKב6q~# [߇:V<(0*FwdHki,8E"qYĂXBT1AwT@vzO=_ؔ&1w %b_3қ"6b*^E bR}`hVki(6)%9Q+ZlxH6?e#⭏:^5ͭdVHjQmk-ػFOY% LZN*dqQmWJ"N&@ҜDX|6P)fV.=MƟG~QDo'hH:|U/xđMq Ca~w0!N_@>S ДR:6{%|%1*iB7u*M"䒶wӯ'(qKc3Ul5=oe ޕ:`UwύH_Qev\֤m ǚLR߯AIEI6(푓(F@wgv2::39$uʖ]rw2*ՂkԊ!]M_upGשALH.S)͐ r5fVA.~/UcZx])֝o竡@x,=_sƥs0/ڛ- gNgwqxnW RiAVd[0P[6?8,*at>Z_45i?9[X YP#G5ݠ;r43gq5vI/e02BGe~al+VOfhf zle3yϡ6G/$t16t3OA)ad }V¢0aVd) jf<3RfJy*Y7WU"#T|遰$x]AḋWK+uc诒%ڳHcG9XUm'ӄ:TӇӞ3+Xyu=ij~y{R-Nd]oN\,4J9w5RQzMgU!ژH|_݆Nj9)B%9ZHNQ$K+b/^YO>:ɉgu m5<Ȃ332#eL3ksV7O-ޱsyaV7g.&\DXKbf.HNc$0l VjmjOPW;m}`dTO;yƘz=I[PsډKsI1!_El^IQΓx }3E[  H<_jsfVAS*?=%ೝō?+,g5Q>3s9N@vῗ :. ゼ.~2$-d0>0vLp @!\NQe'OBF" ~dbpn ھb IEM^7 _|k!LW])jR0 0 I iYny`. OCa&Q( -AWǝ1,͉4 Zɉff^hQ6u5/.׿5$c^hu <ν.5 T2?,CRKyz ]>vZMˑAG?n$g(2#,fmFg??sd:[H1ZL۾ o6H9*=t춱{8 % Pp9\ .xwCimN)jr"J c.2`(BBJEUM a2Z(S|( ]%Wl]P&DIThlP"8o"fhm( 4'li4к$&Vr2"iץ6/E/Y>A>F {~k *E,. CQRX/ߏpc>kIrG'q8<{>lpq~NA G/}~&V^#;`]gOA'Tm̲"seQM,reXOypy`KA=ǚՍB9t}ȸ22+27 |8k#g6;r/7'n,mǻ,t SҬ\\tދ0UC N_o9eF 8)q G9?$I4"Em{ v"O.~Y]M-Q>ӌpnYQgOKC(=Yh%|hkaW% G A}Z8$B%h),oZY5j7}0IHTjQ Ǣjϡ<6`h"T\KpJFmПz34s](&yr/2] 턶R8/[2 Ԕ&>7DL++9{ngN2|>r`RcίR Zk<'!zD4z(5S+4$:F{%*. j2t[V"b"Ld!N Ȥ7!܆L h <2/, UNN1A>h?DՉ7{&_xvUBQZ2SZ"jLx*.A@ppU8aX(AGi6d(e  \;8] ^)q| \)k 1jT,5+cI'X$fCp>SZƙ!|B#SScrmR !AhjJSW VIKI5eWplHToi{-c"5`꺛%Cc`BovB }Zk&AW,zF 3S:X#Xipsyqs?]O/PESkX#2*97':biҜ8`aZ{8hFh0Nڹ; Dzzvm&0f;﹕Q'an'7h@!frs?^EL$X40- 07JOnʿGY<8WM0("4e60A: !X˳x>pxwMMb=2;;4ƀ& 4keb2Myϝ,> u)| uD|FҜo6>GZJo1!n'.31LoevQ)͔i^Bg7f>4yPDnPq;8)nA$T.QLټG+bA@]"A 2)]ɞtשV 墚>|&ڊ2A+[sj=B%-%-*EE?wլjhG|2Mj8]Vkۂ^å<`I>!7wYQY qZe %je mS],𗏅uzEM;~m71ۦ4&=(#202Ό5]5+8xG[FeFsE*u37_ƭ}hz#caþ5}IiJSGpI&#F4[Y5߄m8o| ̝DMd,~#+&N%ӥCљaMDh9&_]=cnB<,#cHOפNohIV<ݶedu]'!'*)"Аp h q*3^L*9+cטWH9ҏluJDfJ dQmf , $dZƺ„w3PcnV#ʽj5Ԙ_G  *j cƜNJjt۲hTA%jHT\tԺaect w$c;z6%ͼqb>ypt`! ?}a9%-)Ýҧϯg Ghoۇ˳MϲOo:Lg !q1%uQ҃n:@5r ^%`l= v F@F]MmI1{6_QCԞY{ ЂXr6/SN\iʼnRfo3_/gR,[Ddoer K+{M A70Yw^Q{a&gH$S6YUg =i0 r҈SOvb3T$:mCP Üy$W_ɨ\q 7M33%b%U0dKKf 4GF4%']0Ɠg&y2@O R(k3 kH5eNd^=Yl2Ym I-AVjho\{,oB4[/*9jDWwݚYW!,Ipp3pCn96| Fc{SKGt.7_p_ pZNc(r'M2/CIT9@8P$O rSYPG<妻"(WӲNN;-;W,r ɻ 4Q MyR(hLB.*SZPμw6}QM}_@sɄB1ѿmrI%sgP䃪fQ.vY)XPlrfʘ+Pn -9|Yse r-O16OATm1ښe ͧ[u, urlq{1#gVW.ϒ;}t_hb/C-+[.WN6mߪ~7rjQV$_6. SFnIN£CYJ%WqYF`&Dp$> C]w2`yiA'm <%P{9ݯz׺aU]UKX_}xIe-O) e){q)y{|s@EoW?eGNmC,h^|<~L^}n[K^;˿)WIi;0(9GƢiޕQ|UƍtEn+܅ʼ2u⁕$2W\zs\i^樅Td"IQ1"̄)9) iTxG |4Lb%\8X%Юyhfkbd! ېү~n-2^ X)=]Ucu.}|us}V/Mjk7?{{|99O N>ޚx-J{lՔ Z֧/Иiͅ4>w8r6 7%žo.BMNk{y</;ea@ډreRqͩ\nML<}O8km^V;cXrI„bM+Nox+36xG/:[% X1x(aW:*-TjHejjqxiDv4j}qWip-ӺҠ%,^K~fDf`cpJXf1(V Ξ=҇3-8-؂b N붠^ɜ6YP9U19M%Xƌ) bS課̼l]Ej=!)Y7pf>˳kgPQؒT|#@"ޥ`.K5a>nSeDXeK脆7Y6Us]<&v/F-5V|Yj[ڇ/]6n\?z.':6JBP)fk1~Lbq+ A8C/>.kWK"{/z8\Z Zo(?Q+oDoMExP/[wo{#tmC'^mC'RG*ޏTiuA$"m#Vk:P']A\ HGƉ̠p*rJ bႂ'"N9T+,{kfo4UYbI A E@ Q%K{|\1|rK,EQhmu>('JyJ1e:`t)Rd9ImwYI̬܆n$|#Cogk};+te HE2l}].>@ϳ;oa@IPyۣp2=*(@{`ۮ ~9:|ЖxliMbZfN..ՏXzǻT& p |<؏tdI#|ޟnd-6P̴H)`j:Aa`EHqx+X3RZb<pOzno,ݷv{K՗r߇[p `[Jk`q{:r\|{tQ^#G{!(J,LzN# P1lQRA䜂!hb>w+I0ZķR#p[}Z4̿ ~ 狈hvY>dn!3VrH҂D~`O#hE~4V8yߍ 0 ԂnڗnuŐsnCoY%䒍6b[nhhVGY CdQ8K2P+sԔdB1Il7bQȸ9zTF[nŒJrq`%vda]LM;IN td/ljPΓ#vBX=u0 FՒdW~}sHX\T/;UTGErտtZ{{|s-J @oW?UxL~Wwqvl|<~Ӭmޖꃾx8^xSmEq Ok,7TN`6z=?_&][s;+L$Su^6ݧٚgT[djeb7l<[ @p>u~Hq=G+wӁ'HhN~q~t^XkYOs[OLchO9:{u\Mhre:knSF=Н/_ںe3jА\E$y=T"X`哆23]\tu ~(|Iyܜm3{6k߮~EyZ{ ,ևc,@ld3Zϖ~hD + psL|XgBNөT Yύ ;ѓ̇[9$ 4ȯpBBDT, J` +5sJ/tQ`o@%{'X'˝d(l 0r*@~>^/n_!Z32q8Ey=s[ocˏQ/?F8^;F1y%-[ µm9fv [j<)`?"7.xC,^8??лcZ)yT?oNz!f/>X9W2]P*&()* [ % x,ޔ[Zh*qP 1``r \,eV6Xyq(8߲VgW`Jbey5"(=8.0–ke}hTZ#^kh b,VxnYKii[>HE gfԙ$ mPۀz}&ew}n"I|C1*T֤@`k?8vgYe5BL%-81Ӳoo$hL_wi|1sLڧIސvGΓ')ܞgļe=9d~>$!_uNjCz%  Ww]d4ZΫ]XKw'tm$:WHSqE'uX G`A ~ Q+X-UZxUqUkh -|\MOR=ey f 2݌ƴuaBM+,i14R[D4 0pC%yۓOD@9Dw03v1Z}읧2R@k1r}0 81 lK(+ǠYSdbE <Vf*^Ryhs0Xɽ '\ f1@JֶX͠: @_)A**qbi VahýgQyzDD\dP Nc7ϣJ,BArW Q(Oـ ]ˠٱ\I JX&8ckSHJ5JD: C(,C`#g+!lv;ڋ^N.W0^H/9 !*J w!\Xk#aAM$V naa`yr"=PkdvnlK6[H$)PBBLb)IbWQ=A$X`~zb|&p#Iou gf>OcUBnPEp ^4Q1^:3hZRΠ)3(M,"Qk|O,9E;ҵ,j`O1;fXKh+SIgl4/V Eu~jSj=WCDQ$S=aU31K{uSr>KvKW3M>oEXOWӫUE/WfVW4\哫~:Uj_f5S= U3O ;O~)@x\T8NW6.&ir%H~ع/d뒞"h ##Z#'BZ\NPsKSROS=ȕȱ]z rPd&g?L$\tZj%>y͕A 5yΦL5Ƽ?c^А\EtS&X&Ohre:kn^S`լ[~uCC~rҩNʾzOtMhre:knSFu5߮+Ѻ!?Tbc4A h9rakZ֔OTJ;`{DŽqLHxf~޼m\"< sD#~j?S+0cx-CArWfa$7#@23C"r{Wfa$ u(2$cB^r@!Nar {Ƌt&MƋH7H;S52; w' fr/Rwg06wgάyiȺߺޕ>Lr8m,Btyы3rs.GɄ𳓿6Wh*7LslY&,*M>\mL?L]#5'c,zXER#!- F 0Ynі5-fjr 6ALbxQtavN2C † ;%/%/谑JQFǫ M$xcNVg!D8߶R -o5k7Sm[B((Gz12kf΂o|8HZKXOԡQ^6( r~&>Os8jgf6 Zqc5}?[m L:SK2Xc0q0aV+u1Y|{ʼn#)eM@=M^4$6@]4߯,6VXS^6A1m;9k.L 7\sX|;bB ! -89N`\|jsQ tSp]MqiaP F@`OcHa=m1q䚪MBM$\K(ߠ8"Vj mYPv \ӵ#y!T|G zc28>0vp\|&zUERHIKQu()j Pܢ+ ߲P3p4̔ =X{$! MK\}8BkE!iS2Se{_bX'ROv A@ "t g:w$icwRPѧAj{{ a2pӣ@JhV(~A@9#8hp 쀷j)\ׁqHQsN5Le, cxp*8xgeT;N9-<(UEт+p$V6Wמ^^~^@qZ"U6q~m v0oZgˋy~Ƿm˵n[6.KRFns2mF\JkFQuHL@Beۄ *P A ΥeUVFiFnɭ^v巪ޤDVڰ[* S]᐀.H*RXf[&weh%C\[%kf̵qQuKXop5҈=rCIhnYW`>kM?(R```xN#aep#Ulj2d٨=\}z3Gcm@k=,XdEC(|qKG3]P.A-̆{?`QnҔ iӝie<$$.RrQ9yM ^|~/B xYc.@FWRIٟi")fiNo3.VRBCmLI1Q)~ȵRRxHm&r1Ҷm%-C[ָwP~p ; kzڜ|XgOcҙ4*qşn_(am(D{w)EkSpvvqc7.8Vo?Zw{nNcRb9|zt(|Ϣ3糛hȧ#) C GG_B0g ѿ/w H`|wxƸi@G(-XɽI:F2U%XN"K/>] '@QCF\*c-qu>[ )rj0X*W˷; /Ax /IKǔ ƍ G@E Fn>Mָx@z"WJ#w;cTޙY:,Qh@0(.fSwׅv9"i੣-ڵc `caYHo,bײ5uC vr7 / κ-x9RY&\38 W{/vQSlߖPSB`w9̉[*-* ?nI!^6 vG@'5k'iup$ͥing;x azp#I)|r'&>jza30QBuVIKdnRYJ΄3}rfs?SF1wle4"JV} CXdCSY6}5aʯ$j-sIsCbqq؄s3VKO/VkܛľcSa9ZF'LI_S(90p "qi1E &h))nA;S"ڳC &Wv/NM:-Ӏ0v&X`|CZQO]Wkκ)$UbAu {PUK(lQ&k>~:y(m8H46\fy *\}% COm4LͰ:k;wS -ecDze't馌i<=h{KS1~Çޕ$W!wPybݖ66,YSMQ2IIRQXLRFYUYqd_32eĥsK+&*0A+5%tq:9'l8 69)e1 k#Yq!T[>V|DO[sΡ':[*,k4B[KYBK8~ Ά%"Kk Re޺==_9Þk"IɁXÅiEIYPG &5+3Hu]Ō2X2TL* l6e\w1՚ƫUF Wj V5xt| X|0@1hVw1biL6%w9\, ZjUgmzwjԑ#El`JV-E3CL%/ QTIDf<c*qKIU^`p/^>ISl H$VDRV0jJ &OMOƸ %.DVxyq0ШA k^Ć:iD]Qk[d rEz2]G͇էm%91F3gRPƫA^!D# i\)n鸖Fzߤp0?'K-+yoLxu ҷ8yXVT:Zg2l@!Jg ZQت6 WY0D^YȾL ~@jUub>)Lh<=Ibe9l(4P]jALMWp'E*-/r,p_e 6l!p P 96oȦUA_^7{uӌt|k $,nM?˔Beo[߄V)Q44w0 3JZg[3{j5 FH dTPV:o; oYOƸ/j.wTnj$\SPS(ah,*F ]͔f8c*E|v}D ,Y ػ!STӞk^..o *]vZeOٷV}kl,P57$&$}k8[ZevMK}jUƚSˤ׍'WT} G N $20vGǟ ;[sP(_>}#ŷhk@HM'rQI$Z;VR^a` z,Ԡ%ET=c$V%M^=/yRN`fQN`(W`ɍXa2HQ1"˦^`WxF<2aŕo NhA:^ )xq7{AAB+XSV|`1Ư.fקkKKT"-sopĿ9_I,Ľ^ǫXY\Bj6[jɸ _W&lSX5 PX{DltYȯr2 (?-|Jk<b+&$Mi0WxŸOGy]~uK7՛֪7ٵ̮iK@vo\b&!C\䚡uykk\R뱇Д^fđ#%RDno$Yt)1yaEح~F GwCK3`Zp1X Fn|aE@}{9zKb@5` ]`<TQW٠V")%Udð{79sPP? )HK$u~δъ98.{ÁAH#-覜RZԚmJG{(p OzZoN2B?Nn0)^Z{,XeSp5ql< 'LLB1"(btp<5> V!B`, ¨ȑ.Q?Mh_̛VMM+taj:P?2ap@3`NV),4fktJ~鎎Z]x՜(-/ٻxGx=Dx,8'b%_D @FH Cndd yJ' f4 +^cV3^n.)5VT5_2>Jc<1 7V$9Bٯ.%QK}\'![h("0,)C)8v S?V%Ur\ X講(W"Ah^f/Wrc…RIr%dN0Zu$+kX!P}ߚqa; vOoo*ीFap1ip%߼z"*]JŠ/qkknviϻwW.2%!Z, o@Vy6{œdA(c8ïGM`BPm@DHO`*&aH (7<]  .1#;I~![kBI}4&*Oڹ)ȋ>+Y6u 9 2 jOV'y= "&῟ojoY) ZrR0DBPE7cbR~6]2߿]32 LHD($J1p)$C\e7T +]ތWXfa^0b\jW"Uh=W[EqLotTD x?9{[!>/=;tzdk;ho9`IO1^=n;9]qmpy۽?uvz2y|ǫλ?^~=zllw~.\=mt篺wѹ;=kaF u7xvdm|%*,BW[2ŧ3Hnώpziҝٛ4ڿ9blN/ioQO]}<5?_ G}0$t)/_톞:TƓ'o8)?X G`NzIl>iDc.^!t5Vʶ\ګn ][sI+ =ٝ=u8/x6NX&u f'%٭q_ԶaR_0ׅ_QϿLˆw]|8<,ȇIH !I` @ە=1ccC>(W<=:u m!3 1#MpgWjqv- nΦrQh2s\[_'2OT%gxO6+9R׽p1;)˜p BsbPT@)gyqpϣ`<9 ~ΓK¿9:G3}_+p0_mܹJf<N960c9q^a A35i@`:V(w@ M]; v@p@ƍ`p6,Md)m"+Md%aƁt2&{GZy/IZFސ9KrI:X#j ,3pUIڝN޼Eu9K[-U# vzKtl(!lw:V{96_ZDQ,--)[ZZ]'ʰCT:os*>?nz9>|b H|]=Y/ѿ5O=(VY*V )S9f#6Z`š("-7{C0{I:M"w!Lh֊WZP" ,91a܉1۷ow8![V\1b [)JXHJGʇZἣ.ި́kTh̥1Xp @IzC8--4"fP4e\NK-w*AߦǷpLi襖B7[\VYTWPGU kfp?`W箷DjS2K>9Bx7D߯;'oW~~7Xx+͋aZT]5q\+%V孌 !Xokї&"Qc ,"eZ`$+#UH9"m:Oj/l>;d~cήh]̲2#qԴL?Áe>/)K'2>LLm] 86`"2,<3ȠLbC4lY ԝ5 oI"sPfҹ#^6 dc<5w7SyH@Xw0pX0;I\tqN ,Kl\-Y]u3lP7/}hGfKKRIfi VwTh++;RX.zpMB{6/]p!=p'$zM-=ThY-.E\чR hܭ` js0"s! 0hP޶}AKW04RTu~rsUd Nݕoub7lx !Bhؖ6ΙRG@;=@LMKK V[ZZ]`Nt,$9˅HYyzA/>͓ԟ-ݬ[~ρS{|8vy8x 2?{Fnj>ӊADS~IhDă,vJ(qEDFi~TJDU*?_f]}Wxe#d,r,wL0IT-)R1iJ2&OzmR%+Gťځ͉ :-U>x1\?~0w:C*<2:vp`$g(.˥7Sw NƅBLC$_/_׹r.|J$ >BXݩZļ05# Rvp%npe,Nϫ(Rg&]?<ޟxoLW2Z(q޻9^1V' J"x~x`J ,ug)Q#IGi!0W(Ƒ`T ,i"v&NRR4"!r'ZN= 6vQtA]~Pu&L51@(g< 6ZANkAf0ioM> 챥whU!gbT)_$Y7&:{Z4l[n2-3Ys!W >yr_kFoO{3qi+s"%D`g̩RTˆ1uOk)< N `M,xܠ?#Q{ٟ=~EMqfLfIyfpN6)1P2^eL)@#@VMʠ8` >cb@,9-`Ku!ЌFrCMMzןݕjQCdh}=+CZ(El ݽEmxZ.-x$NV}ĘbI']r{}Vz r)d:ն7l=eK ЌL&#AZH⒂'ќ OOYibb& % jQȻcK4K0 =O[˚j`R`*^0ajR#%[L1`YZI.ީm&>XUĒ@yM9JIVY,Pj LpnG,RɩBJ\%%vJ B+gR*g\ @*rHQBoK1"rAލXc KoG."k$SP }G2anaۗL@Bu&dJ ,^ED ufo>x%%oų:2ӛѝwqfH{fɢ`zNE*p:EK9V1X0t8 Bќr-J;N3 {$+u_Gj Z7V҆Č>kYTggzT1l )S2TLqL1i'䢔k,Q4$Ei%lQĄelLji)wlLuqJҒ6dtqx 9,ef)^e<1Ya,Ä[;C J!0!'_OK~*kF/L( . &@Gb,ݺ-Gq`Gr?xYêSd3m6 Z}xtbJϟWFlvgK)C DN^,@!̟1${'}DM1Z“"jSZNN]u >uH'R*A AQQgئf\+Sq["ĢV8`ƫzH>HB.W'o6#.۾<],;H&]LV2'! ߤkejb$`ʡ_W& E@ "+*WA 4re]p̦KOn&ZS,%"y Qˊk!]+TR-NVN19U7|xmT ³́lKdԍW 4˃ ɤ[i: -.d"g(|NBuK^#*\,Pٷ 9[hgsIf)3mR9-"XV/m[ڈ{GN_1mdMHĈطT&^ɱz{u,H;g{Ԋ>Hl=)Hz%IRQzӒZ^>XY/K-$g!ЏgLa n/"ـ91J!d=ô[w@ Aԅ0>YCxŦ 2@6tWt6c-&'P2k]i{ s2k tB,TTNJv(&;b*apb p%# ys~ Z]N3{a;dxn %qAQD(5BNf zm35D'Xl%\< 66({PdX2 cʪi?UZ=A2 5w}SxQ_6Iv ګ{x:gJx/&vU!O]dF;60Q|#P}N~jՆ9L?cՌsn -=!1G}) r\sb6;I1A l:"nB2ziQ ӧ-N~։@_?510.E' &G͋OY}]ceGΫޜgl;X;>'^}$ j:3i=_a ƫf'|r!"Fr]X̘ ~ę L!EqTayޞFY»K<L/˩;ȹ'V:R"CW-z;=!L((ځok}{^\C:9.ۓr(g۳.ju=-޶Oޡꮼ*d-~wsꇣ?{QyXK|NWO%t} ,U\Ov"b6zh7~zo_d^)`s>f[;yܵ2Rvv\ۨ2 a$ۙG-_m!>C5}19*6?9✎H;¢JWG) !b`]j(U1%`[t=C& ;>̊6f,a SXRBY[Ϫ*R`VuaJͪp- A"G.ɒf*Rk1 3FM C `a}$ײUr*IǔF\%q xsi!s >kE?isfNZץsW܊!DYൢ5uedP< \de\8R&"( L֪"S"d-JcIIڞ.Yk28"wi]%Z0yxh~/U SO#!%=^c%k#% U=9Pf0?hQVJ֐^Ҙ)wAptv7D!R!Z"/|ߊ$17DoہrV}x\zI c`hީQnZ3BbeՊE%$k-q\paW9q[n}Qn8.>qy޾F5yrfeV-K1LRzdw&Y4'׵^\צ/ۀ3%QXt0rcpt80v&r,.ۨmѶKD?:t2 ,GQ{<-$s#R.$^ ^f!'{ +l1ryi6o;\-jB:.L\^ʏQ>ņm!1Y%Rzv]!f,[m^bėy5GH Y[T{)5B8,=c˥ F-~cPZ#Sf3#έf)AU j4N:2(;bdDlSvlb+Uh)L,it)ddqUY8N!;fO\9@FD*"q*tI2%2W Ua})1vӳkyHV˃f╽rojG겔P!ISˤd!JQZJ eFȔn.9,]i!8HZPO>smBm'Nȝr;MAwAL>Zr;6$,0(ʩJ6l(LHUgjv\)ڞ!)1| $= C~[7]sܝ7Yt%TW=z[spk2^ 3S9]w!pŇ v̋|s|mWau^~^lקQ?JwV[vSy-ſ߮-»U8_:|]3=.$8pȗցzݤͤKX۩޼yw՝~7yګ렁ݤKd9R>=FKڨO93Q4wځg% L={AU_}(馩o}uSxYG}釣[ꟿ??4^\\0{տ~So۸Xjcwid9l?1 q2rW޽ZI,~xBʱ[Q.uZzu`H CN.gڭ@=vO/n۽4V=n$AHi?Vβjc)Ӱj?ZsZϴgW'Tz;}7K^Y͒mޜ%[H'uJE*!fO%Цb%I VnUMF{绵Ɤ}zuy דIj['䭴7FyUO3WoC>~:ɳ>ŬYzm_JvZbĶ6Uš`)梪v]JqQ}@En1R{kOth.!xo{{{!x23(2$I$ )N7 %q! ERlWZ!xu;˭ |Fw = (~鬮=% §8ܷeA/CzzC~+"qE IbHB1%SBL(Q#(S'81Vu;v7>070mt<˒d,÷x\@I)eb_jˣ;'RX'm:G%(%R ȫU2^/jP xX^=,j jFH-z%Sp|Xip]m;p}U{aVR}QYUKJz'@&ǡۧv+? ch;8G񖠡h} \лuei*4zRRI\m{t1ؕ'gi*]s.]JZc Ll#o<B7MY`NpKɝkl!x췬QWzU-Bby꽇~28 b>a.5 O ; KVviHBw(ISJ'J#QɝH%)vO} zhY"\K*`NLYB8ڹ>r@ٛ?`ض4JY~Cg690"2;$#e3@"c.2~S`̘zܘ?K2B/#"Jh%ze ?PԆDj$FyLGW"0_a<_9$.PM:ϕk:`_Aigp'OSDIMʨqeăl0*8>a^gBc-dCY@>9l&ܻ@ w0hOnrmPoZS4y ?fF t)ȫ 0}2,qDK^b(6+[BԳ|n+ `}(psVR/ě7x6z^)JN>EJ'A\`쪕|;LS`E AHH*4ao!9ŤBŪ \3V/@됹ވö oUtm]Hx귂 (UH $dZoH/坾/LQ.R$8WjFn6-Y ־41?Dh)ӥP섍vn= |D,wY@9vXʑg3i=GUQ:|T @zoO!ެ޵Sp~9t;m~`YpH`R6v(poNzu :꽵R 8DRG@ZW0wە?{|4XpD)C]WUKݪd.=5{r_l CCH $$@= UDj%Ae6 x%Ƭb\ZS u$XĐ!e@B"KĘD.f<O,ԟbzYeVG[ FdZ0 ,2%[jHO-4jy*O#C >Ĉ|[8|a l>IT UHg6&nwzpyש-?[l0[8T6Dh*y4"R2:a]NK~Ewj ~||"ວ܋@(қ8nm!Pޭ"H/a]#N9ƔBḗF _nnQdzFVar^S/*Ġtd^r񾂣b.m_m.oϨDn-$+GH 4D ';f y\AJ QV(S B)) 4"kn}0ip}=j❯ce}|!dl9 +k/} zwU7ή5DJIF;噗+ybzq¶2GAqš`H/giS Q. d()LHhr2Til d8?)s_(^_:e:˝PŤa9#P")8#E_W NL$$ 2bq%}"sͤBfwS 9zXׁ4=%sa)j(2XPk׉ȭk0Ͼ<t}٤ί."SLsנ/TeQOw%֔9aOI4%UcӅr~Rj\\PNhM:V59h[] u l dp^&ttnȠ̢i`!DmEB*q`&0@oFB|`tAc>IbAB1@rJ?#F`j a7 6!$EC ˎYju3, Qdy&lZmL 5L AU!b*mZ&CRi}$/ljTۦ ocۆ3H !(_̝FT:b9O鈉*飵j 6Թ63G1acDXsΓh^\ਿ\b?{_\J8.V(+I랮HJ!ANWr =\#ˤh&uWkɘ~'ҕ!l* Ec7 \k;k!u!!L5ěk9k>˫) IL@GmRư^]MoiU9 LZ-R ~ +yUedCaY=y&!΀/V i.qVBHPuX + lW  mڅ[ |,J$ytɜ]t9H\!OP?Qm|7OIXdqПNGsrŘqxne*qr`5B+6q ք6i ޽+pu!Fe61)2.9> bsroS\\X}]<|PFM@|5,Y:_&Zۧ0%t{t4çQ\i,C[W5avQ_FjWKy4(}/jx~Y~Hj [ PRp4JFj?_~NK.O?s?>756Bd<„R>FY@VG΂Hץ #-Z1 [^ t2VL4AM.Pn~QeY>ܷ뇼7lGezVH[qgjI=ODU뱆zR}4O`[_=bw#D(I(FPa(X FY8RbI4w\(/Xkk Z 9ۈYSYgձV(Z gJ^nF=_[X=3ͅdU7.B6t;PӿRiׂ@v9;J)ъkR55oflR0BMl&bz!\VP D[?>-uhhRw7'oy%lv?=ήIzu]8S잤XOC͕y vN]_~oj~Ku9(#T_'WlvLҗl: V@/N@L}@\qd-^Cd4-r!%w0Z|op>ݿ*̟/WA̽8oDooE|M>E營,;Y5>Voj4zyTɶ7+|N0IfJU7ǏߪjSL+h>PoL[I!XGUw8}|ϳofeJ,~tY Gա 7sUp%kǿ%$0C21 i&߉Y~3|H4ޕqdۿBCKF-վÖ3xbKj9Hd{w)5$%EDU:ֹ]~߲e}K(U~ E_ew_a t;ߛyϟ̵Oߟ x8~Ӽ;-@(^8և.@/4k@rU(o&J2kk|KgfpVtLiD7W T؟ltwBw+[1S\}v7iއճdZ :ɷN}o@0 I*0^ϋ܍olo`?.y_@>{-4qO/8~{O w ?_0B?[5u2)z iI.ȳNuXX/tHWu@ ʯc:,GEH1t}4=yg:QT'|?wߏ@Bo ^nr~-E.tpb,jח)s ir;Iԭ+w x`SqRӕ{s;R\L5Ϥ*cZYȂSf&lA"tt swWW p:krEQ(~s߇1c$\CS5X_97|,..AW[#(RQWv< VjwJ.ˏX ZVHgr3<*AvuA;Ã(MAՒ(zHaLJmsGuXh2 B X ԙ#F4I0R E`eLF% #)kY4i1*7 'k)^"TE@IP,L74ɕ oI:T7\mM65jTHGңkR \fLNME=h ٗG=`>™wM֒ zP*!S q!ךnQJ&9kk)*L+T~_:z̜v̏Y~th̉gWЖGП/aKCUl1vn 5^P}}IldaJ scRpFE@^C\[) }.^ 89A~p&cj)1^۝*i[]RֽFmOM5jI5|zӨ2Θ-c4Qj^4|j0ߡAej%s44ikũG[)FE=;!޿jYG(}`eqх1Ji%Uz*Ik.7u0:[UX4:EeM J4zw7Fn9qՈSQ8]w\UT`j3{7V8lG6Z|sJrQwhb'~´'Zjz"JqN3j S)'V W @#%g 1VxQRyRh;'.\ʊ #9"Q WQ!R$HJr RAV+V@ c7*[W9k4U"ƅ [j FA R*KI#@'Ÿfh6W :^|(3++IcZ u $Q "〬1F3j#""X:)x R 2$2*#B0dSU4հ)0~LQi # !Bh\kQ !N XM% νwhh\ׂy,k,lG^-SEukin%,r&%&%&%&d%jr:I~h+V_Xkg_i6*\8 :˃`_S̉S" &@Fw/[9 NkN錱]QjB6\;o6:%@j)EZȠ-yͭDCB6Bs:l,p bp :cdfPgnTc0` 'V8ØNi(Fyl2;KSXDؗ^^}k[YLqL[aY*U 'UU >(;)y<c HFE`i &p ZF|.i6,p]d40)i<ySt>#Xs8޿~3I}RO{XXkEi<$=>KG{P.!MKyi;j{ GcMmwoāדa)׍` R `Ba3oYE TT_K)B A~wBZ`҃J8 ϧG /0$'A7U»(4'{5jd}CywSQj;:3`Ji`,] hht<icfJ%**,z̼T>[|BM#t1 Oʨk"遲2͙ɢ 4JXhIT82Tċ5yEhuPk"FDO1eK Hrs`璮`ĝ ЇF aʼn^ɰA/% !V0lZn/ mj&.JT y5 [L7(uVfP1* P_¥1T?JNX b@O'`!;pP# & (hӨLXڄI!ȨS&cfs0rI̍WڶG`QϤ- cp/3IQ|: ђ!)f/_>p򧻋)9K(9e1&flY漛w)Vk7 n^ʓb8FCFL@>7ߩg!_S~:m{8:64_}8{N˻Sj+n*|V,<=`ɲ3ً+.1Pճ|V;WQ[ F1a7^Ctyy!W,M Ay:_.l<^~Zv%@;b>͖9mJ3eLľ%xʭe;*d$0&*hƘv+%\M?Jfl;ߩf OU-gmƹQkƜB*F)[j:wV)*1+Wl#)Xez̘'TS%ِp?Q&)C٩r;lc.+*Q@];Z&&,Ow6)V8w eo 0qAo (mԕ4$ǯ(6dv3[5ԢZs8^wqyN$k}G`o[[PFJ`)׬]{gu`^:v]Z6uV{o~Ƨ d;+NѶF`7V$ӞZI ȯ|~}W=;O=`07 R5eJx˴1f1okʮ^Gt7eUmP9@;V*Ŷ8qhkɘf3Pgdzjkq{"j;t5ӁqٴW>pqީ+\(a7hga͋vQ 9QDz:*ҵO\Jb#a~|Wp"NGuAÔ,A%ɴ!C6s:ؕCU^gT#^VI'7G _r; uo|:ґyTВ<6c>fn0yTeNY-U4݆CEDӆlc2&N,Gv%s #-jB)LHE DZ&pG Em Sx> Q÷}5R}hS+)$>N,tvimdNA=VF XK ɚZ8ƣ/{Nz)Xw:P37ԥPK|0wɧu9 ]nˤr=VUij$.s7+~^k,HξȲ~Gn[z1چr>{i(;n:| ZZ:`V}|g_? S&OziΤT??~7 E{ۃtq¿O7|?s]%[Bw]5m.+m#9Ї}H}Wx-fDѧń"$e#3CjxY3DT?]U]]]/%ҩKm/WV^:}\9(/=<*S[uɣȫL;|PBTݯAStwJ.tg>op8}׿OOR;S/].ܦI檤Z_8r>Np`n~enlv(7]ϙϔk迏uoDIb*@}QJPjoTLܸ=Q0,p3I838\gHT@){GC[ 5}*tNjuF( &ӓG:,Ml\Z:pҝ`ytF4My{r;.;LX~-4۴)~s>XwvAo4Ň30oo>~xQv؟m2 ̷T']y``Ϯ0C]ꬖrp?-cp0*ΰ }T"z3yR¬+c!\Ԩ_0s^s{:hF}ߏ@$$L2ןfV?-]}a:6u{[g<+}zūO/AޓKs?g ջDho'_ޏGVH^CUoim &w:&xts]te([G_'e*PP.yWq)zaeO.Y| ;"Q>l捆l~{KGxPlt[`R>JH̊EY塒EYxRaJ<z\]+ar.+?>X L7ڞ7+Fy€Wp Wr2)҇3BUppfb04GO|҃sI;|L7!빦ks]MppEpp7} Yї0}כPVOwrBcG6`L1/HȠ mm±d@(Zul>C1zP^:W>,8:hkh0,`m/JM|c{YO;= x`PȫG$(Twkmujzq;\g(Ύ#ӭgq 9,m4"30#V&Ʒ֨pu/CBdϭG:PKҙ[qZr$keе}rzͲtG6 e"0T n#[qۆ"nęHW6ܑ<lV;a^ EjpL,1EmnûR>Q )ň9DTK fj{)Um/TrTs\=t'0${1a4 )w Fyt`SM_h˷fche6!AH6؋A0=GsҙE.]峣<WߞN+ɓJNw5hiT dS tq2X,M9k*MdMGHPZNX3qtF3a+|Z%Ҕ(Mђ .BNZ]D)LPAa80jMa1V9nJiW$:Ɠ:T>qS^ᴬ%Eݪ҈s7ҟIsX]i iW՟Lnepz;b7cn_Gg K +B.lWiɶ?]u[C֋)7Cc!-g Ĕ*\ZYXӧ=v(tqIJxD7OՁ'j88:)n^X[v9 7R,wQ芀Ix˒54ȪR 9/;T&:d Dc9b'Ksy&Ls:uT}ue5ԸuONm|zt NFP `:l~YI[RjWh<8qTD?3Ux;5j/ns1?/WAƥtIo+v`u|+vENjF@Rvs2voNJ,kAop|iq8]KkKSi#ZɒYM\s#í|v+Q7O|`|pkE 4AI80-W*0"ִ#M &sN`)1q:16[66:`2\=CĴ C+t`ƉPqRaR2kZ`e%<t  ;E$y`^@$|":f_DbMm3nH'G=Bd #㠅 #ƽDei䰓&M8HD1b aɓyŜ:%!59 ʑ+b2鰡#X!j;Z5Mk oGunMzz/ZMJjаǛP 7ZCay%xC84 2"; ]f3atM%Lykx#u>!{x9.*JaL8GE<}ŕ_?89H,/d0y\^ZMҖ^c!k7xhgK [g^$M{x摹뉹U%k~]Yntg9*| ͐ T.T2Gnu1:uQEѩCNsn+xj&fȅOFNRoo ҂(L0Z>(JP$ f@T `vT8Zo@b^tl%-TF43-׽ՐtRa`JH/9sY&@L)HLV1v5 Aֈ[))5ԒƐcNx ue՘rz@9%{])IC`!D LV*(aqЊ0M)oFvV\of:ŧTYFh&,k*of[410!jKP\糪)Z6(QLs}~Y.صv2x~m%gKCB 1qZUN.H-}xM)a]YOnH>a=K3K8V gRn^bܡJt4P~|@=?cK+?}5sltI󝬷׭W/Rl6<(PaI*zDFd5P]*d.?XT3J^rk^Yy9gwz"X粱kQ;-xfArW ,8O܉jhaq%tԫm]bĞ3q +No^]8_kE Ҥ"Hh%AF"%*Bƨ-AZ>g':Z.ɨz׋AgK;*?? ًk?QPu67)^;ۃ=5P3iS4tߪT- X5J*Tk<}$:򔋖Îp׹4%%djyT\nUq0޽{"݇a{ rF}ĵg[_CB􄞈Δ!Lt⿟)ZxjcSZ7.MJi8 Qk͸lCN츞j L"̏KNp 1պx:zׁ*RC ^#%0O(f.$q x"c BaJ -QWW3XՌB93]>~" U}?ЋF?y$N >Bm:|ԹJPrFp˟綾8y&u"u+4-STɣ>[5C.:^{ k6Zr  sg_GcIu"QpkNIp;<~8|9(5ri ޺FdM.\j&K#g6~O`\-3Nӑ˃Ӈ!z/P"+N %Vd@׃]ͻ˛S`cK!4j,ĶX) /c4?/TSy}W5bsͻ#8gm Q$Jyw?xˮlOqn@ +*!82!8T !NHxR2dlٕk >-%iY3u%N^j}IdXa{w>Ձ>'薭{ѐ!R阙VQ"+Fp!4 J!5BU)yEl:(ZP*Ϳ?,)ϴ$ϭ%$`MDp!v ?P%=u/t=U"ceZjqB4w.uQ(}񃧭բ`ШYƘatVA<;wĻ 2tзxCa |n΁T<Jb\MbR4M+ DrX[eę2;nU%lyrUd :&>> ?j1I[#h/N,4BMԤQQ Đ+F[6^| `6I$A0D$![ `8I=ԸmB Հk. ޺|ZnEA$OVfɓsPFهZ}#P[)a^޿a7W_qɋZ," ۫rhG|9[~2=)C3m=Iq fpɏMxοꄖj))Kpe2܍iD?^-BƟ][4(#6-̶-pѬ +S<tN:|!w '֞OE5,$KgګOEGSEևO30xaL,%t[]n82v" ìssNzjX9 A˳ hqW&/*F҈hBP,L8VL K5B[WB cB cTiBD)bSM (qJHnO$ԄZG:i\-JL7 #z‰-CkQ#e(F (eiE2BEqZQ&M j!("]ZXM5E`qbCމ=USLȔ* [xlE9 zʓlH0^*vbEa\&*sgNmIlh"ҝ*2IBk.4u%"(r5m 6._5%*W5]V=L|k8R:B@R!]qtXL0K7"rE^qȚU!HQ%yR@ܵS+SXŋk5IeEN$B3(E@g O283ƭ&y9C?, zL<R@R.=Z4h%"^hVې^{kjfã<-4]u,3|ђ4s_#9vq1U [* 4 <Mߗ-5RXОhwlKPh:LP]O&_p)!E/sI_&6f~ B }*3V^Dw*:J ռ!OqP:3`yGlh5 YB&`/2"Ѓ4IJRf UTHŌ%Is;涡RqQq\z0ЙfxLn_?5%/HZ?]-$o/ M=i >|*JKYv]Ը 8'\zs! CGwd(N&#eֺ!w^f8t^KhIU3wv8o@X Ztˊ/pl&+40MlJ d48[*Bj75v`Չ䜰5Ñ0D]( n0Pܝ'І(ih XCgQ*QC6"5yaԵ$v.:2mԸ F qk)ze4RE(O,1 a<<ؖpd <t,gVKRڂHTDQ, \l#mS,*%4|+<%03Zv5HZyr%Ϯ8fp3D.#($&C(LU`iRc-yŹ*SW(4+C  9N'}2>ܔ\< RbeS4c`qqk\hY ~)V)HWR`C@ŞU1f|xsei6#qPrc Й8=hx< q9sb 7;\,sm\Kb2$5#o\QdH֜+rWɓytQ^& >˂6i]fg^ =lf/5nDvlݳׂ}ڜJ˽ @-LDu[- On[i$ F c\ǖ>_^c.Kr:mLW^Vod*gÛy ݫ v !C>O ?]'k~'Byjt";e NB4>o?_{v{aΨ:~BJ]h:|")Qpч>|`P~2>w'lH(] *R}*y֒ܽMO*cJK~2v_Q[P⒏ZJj).2'{e˵!}@: Ƈ NUr+[ /fܼZ&0J~1R;p+ :8u_䲷Zg(w4_G{"wS .ilT{U݇qERH4P"l@2WvzUn{n87QuHػ6$ۣ؈ ~(L$JGRN W=$K왞 Fl͙_uWWu׃2p-`9EI&,ڂZK+K>1Lt_B!rFŷfjXG[Fbw}+yF~_ɾLq\Hgq1/| _WU ֚8p$!9C7 ģkƘHiNq3D5z&eHgeSF qm(9ZWeLPMO^T"jJ#eb}2_,n{Ձn()!_¡ PbO Y pƥ/ ! @aUpL֜a!נYjTq UǓM9^-meJy? yXjZ ?>\rL%1v{ Q-x-o#"`;o\l%$?t ヌ2)4LVwbEn^K>^cظ=Y߁KN)kШhu_t`8=v!CΏd@g~|a=ؒv.V3tBn`z7wwOqu ?[ӽtѫSUso.G3nqǛiT.k|''s#3$DkfmE5wn9=Ъ4R\8iڜ_a;;~4_W"-/iBxOO)r`[لwq_|91%iqRZ-!D!KDa$Dǧ257Z~~f`ѪԃG\jœ Ĥd YbgRIf #$W"T%'y {&x}~B6 c~Mq6qqDua  m ,,X FD"΂NJU&A^Ά" 9Zr&dly̐q ̐q W3CRpͥ1FARo. kR$,ƈgam!J21@c3+8ȁX*0]%#-.n};9;ʲ"gΊ,*g>.ڝ %4"TvgB 5` 5FtxQŵ8=p>:|"(Ձs-hr!(cFנK`:%[@aSu'=RvWccLKŠF0ب{S1*RS1RKR1S<.z]~goő$T0OLd?K-"k\/UpP17d$Z& A!-Cicl"Vi ͠ZDIuTPY>.MA[1"hezHAr"rtOF[H5$1nWIg~j5ׄR5>3LZFrG ;};֐<yL9/}u8ET'I:ܖ}Zqm+6-ye$ [&𶞰ګW3/8┍t?:BKHD%ʸs-hVZ0i [ɾrtOF!$O;DpcaFc-"-yE$  o7dV$Mrts?K#NR0qy{Mlvp?R;B8rYݖIx^LWWYI^ޝHʹ^=6`Z-pz=CNh5> Lյ^rnkP/1He}rqm<& MKGi_O@ae \~%Z_V GFD$('_K*❁jψ3*zzURݺXe0u+ QD9 a׹d;vbr+Ip̿*R?bg=i.zIbE&"/1bew4_OYR cdJA :+A{ kbzWr::iprWkH3tG%G 4f tQljP$nu SZBw<*̄gTf K6Ih8 Њ#ڱ iQcjc]oԽqX$ѽ s?+t "}Xz;6'-mn\/2>~ZH{3^?vuER0sA7^eX]`'h]QSׂ(jXrkf1L/S_M,JD|3="bM,^'(2uRv%jS\5ېu-.Q9m&2G\#UWcw PTBԛ D`'7&"O= 5nEuL,%ʅ{wNmZGdŨi&2G\Uq >"@XT0q7*%)= d~N'pRL ɣh ޘhAmI@6Uq A(@Fø0Evf/ 7ޮP?ƫ9tD- 3䁐(nGwE<?Q|5%f͔Wo}{O[ I jk#80ogp $ &l7|XU'NlHXJ,`MI:VBV Jx\"O(" dOht^1}W7e՗]Ie4h(&VdfQ&/h* _Ck;IiB6w9JasNY;ղy٢sUT~EN (%ǒ[m(E` K1{C}І!%uL7{:TQm`.w9U;̜{P ==HEA4oubiY %Ƙ4`,f'r$ubJ; H(WA~̔EcetR` >_q+ *n3/^崗Nٛw2YAB" $CjP DA Y(?^Ţ2G涝L G097>-D ``FR5)uev04&6p0xv\z t"R BeG<~aAh$"@3#k+ "n4Ws \1Mx^}XUP-U6M-OW3mRf8pGg0/,%O2l?fq ośZ]OGa=Zd!-ӎ3s q]n4Zֿ*Ir 3#RV^n\L=l/vò܌"bKbA2+F?vt+qI|?GX̿\h_oiӶcᄧ3M:T,Fj=.ݯ^K &0D}A9 &y?<"Di.JW& ;t<I^]h)DR/ &8D0F$c&d+I+cT3F;i-}U8Erl OmsN"N4cp9APmȐ}R1g0˒zT\s$B%8=SϐG*˳lr {ÉLKkLJ![cIna 9oưnnFˀ0sg !Z)/20:kd Y!EYѹ&g6\ɿBqߜ`CFjJCp+\MPk+DDa*qoXn N>~}\P ^ﴻI\?S Ї1n\m dO .eoަ3$I;N0>J'*UUDuIeMۅ<,n uoO6'ZZQyVzG=Ȼ>ߧ4'AgH|44+o} Z؋u1݇%Q4a:Щn,Ҫ9ZlT䇇j׫[ގjd?'cpU1ņ `{_D筣04>&Hj$m-Ƶ>fQ Pl-zߠdȵ ]wIdp'?}prTĐ| $ xߛшJ [o:pϘOݤ<){{.T2K-ێo_KU LOEڗp蔶 5*Æ .n2|( De(deMJ={j$B"g2E A2j}Ziri}9^]eYDe=І 3 =D_p)].xuMXhPB-K!X퓥qT\`*]ɠR)RzIa ڰRO#eOIqrmȱ[^L0v:7KɴS4Q7Fb%4er~,S6fJ%WM k;z^/'.&t}Skߜ|(uE|΍o9]Kz;dӂrkWmtl=ʈ׶\1xr; :?$J/@Aa*ǘB X1}H("`>xn`Vv|pb3,4=_֔ۘ $B 6j]@)c"K r. yȗ &X䌸Tf憜=6}n8F~QK |?>j`^]L[!cAm2򲥚'WaQ7o9!{/ًfZmkl*[|شEַQ>9es;K3w}\wd}xUOuryXf. 1.TZ^(ȝ~RA?| iस^ JȖ|t-N~̀ԂnȬY7Fl^ڷڑ3'[,{ffݞqߡERCfֳOodt̺%d0z`ݛYJQM>{cҬDLRl'k:O\Z56ёHM/H!dϞ]6*O`*8FN%-$V$zտ)Dq5RY=Qn"Dt6FUF/FQKJ $>!tUŜb#_zM}>QAۉN7t8qnd:`rk~ 5'Xru2h˞7HF&zmJk;җBVt$NS*L;Ϋ]@ID.{3p谳Z;8a mk[e߭اطjU'*5Hˢ)VsA*u$Jh6TbkʒѲQUep*!Py^T"z#v"JܹV\Bꮻ~s3,'0am 2El ޲b([)h{hV D+efbF 6F+^1M3Z1vFڡF*:!'BCjY8ػdPHpxEi#l h%Y*/lzOZ)6!<.5|,9kvDZI QͺzP^w\jR3hA? ;׈bfKۛ9ZdfmW=h( CYh`TaSFRr&k\z*XV#B9‚bH_3> s<8ޫ+ 5 ̆&"y&ԇ-gD]FUH-X Q.ЈFd:?=S6Ln"]EC<"f}(8JYxKqS+#J NbszdLdv}@hQE.t$*/Ի P'O6QcթWҽ1ٚY ;gTZljh ߬gO^ʡ' .Q҂"z,V1:GVάO_ S,j3M_/˱uF0+z>B:O `x:b+4\p ¶\ |BG)ϝսGFF$oLR}[0dbZԅ1e?aWa5C3RC}JvG/C2;"@݅KGFC0*]ˠ PXϵ (Vq gQ,"G1w'[SAfjMu־N8/e ޜh/ QpVB{JD뽅 =Xdbw`0cB'L9ح[;J>zs(9ƢOŗX?KX}$}lBQ|?T,Crmyp-3g3 }Z?\F7"94:~Dp1K&5Z:pd% .JY}M_uWԯҚ_nʑ}[G7rt\oGS٧='mnRmMMSh;Mn"8 PFt<&|19\֓oVzA{ vf6I×j~91vsʪueK}/𗀥J/@uPX1F;V`O09^s>NN[F[&{c;+gq{r-B#nhy`ҭC~lwb9jAuŷ{.}ߣ< K(~y1zN9 >BKalv+q\魃}J<Q>dZU;s,{jh0 |q&ޒ3kzt 9V3},mzJ.[OsJg993ɾ=+$f=g!~HC0D`p1J݇SPzu-gPd#ⱈX\""f7)i]1r41rMxBHS=s GˊPhRȣ(l2߸Bw|[*95sXȽH}%4!|0Av?9XRieҬ2-X$w$W;րD͕؃Mhd_EcXsKޟ"9%?wf=K׬g,]կY%mx^2W!M{C<^HW$ГIG,Yr }vVm'o!k/v/g),֞ڳzjc kknFݢHR$Tvo[)\m$eo4HIQ!fC.vDpF_F#x&Ӥs,ALS<Ї=EMB< ~Si FV߼ H"r.Sv/nެ^2?~MQMqƁ!zU_b2sr7Ōf u^9֓OWQ1!z M VZEGH' a9r4oRY) DQHǠ Y8oD/I-sYE9)3 >pA-ih8m'֩&$A*zO!5h$o %,}Q9 0rC NPl!`CPE]Z QR#_HM86 '0D[eHitI$=-\[A5dB:;ȐC# d A Rb2 b:6*qR |6?9FFD"7IF."ۥKIΕH6|i@R իjLANJZpd"i l]d e<"j˓tW &p :2'PP7"H%UA*_J/7׮pه/ubr}q5wO{%W=44V}sNˣ_&W/z[\富~_Ҕ^?^\_{>}΅[v*~-n?}["ᄌ|bΔ -nCqC)tmG֌)F#bC=B)Nɑm +Q_z[EMRChnaX#6i]mzAʂFH2:RH0!52!YHT6GaYKA;PHE=t6{z}p|"JU$Ey#iQ|Pp߈냚fNmF*2,/$+Px@(蠇Bb :ldFG"MBOs&IƲQ@ dds#˱F;^v5KkXr/XfOv~*Y[x5@^׼rhr׋F͢6OD &sQbzw<[y+̰nGڻvhŨjc9 z2Q2@9C̎|],춽`` =s4Un7?Y1.pJk(%zޖݩja9(ʤ#ZE㶣{޼k諐vu/w}Օ}gjz'~47(C/ Mjԟ?|4 dIFYΣTO*W{7}7g1#;iW_Ӱ gt2?9N1X%3g9j9y`8}5h?qo@Sd5C4ngD&C_3uun3RO~⨩ͣk\R{t0+?i_%}un~}C#/o\@q/YJAXLR+sZ&N?ٗ6q&$MS ODuWiKPP #]""ӆfRUNh9U14apY 0 q-(D ɐIU;bHVHT"x VMfTo],y%1 U{Wss*M/,$o=0aRT)FlfANAԋ,hrn"gGzx w7]90r7@JV&8$d'xe1z5GG2Gz(^Fa4 tVIY-6I O\,Abkw /I 5!j-갢8zM YTzU㼔Cvꏟϴg_nΘ堞P׫RT ZW7PF,77x@⓪z!mG2NnJ#醄zg5zO[p|oQ 12z|XzUyW,qCb[3c$FN2mx5òT X|$qK9҇)V@/ 88Pi"P0vdI=Q|>pڧqB̞JT왔 n0@ppu2Cᚼ) SJI&bG} QG28lzt}jp1DzoQiASj aK 9z@{(A9"Ơ`" b֋d&٣ЙNg\ZJ&FF%\ #b"ZГ#2f3f"$keHN]"x!&Ew&o+n^(_bQ-'vVttWDJ/:tS"d/7dOί}ڟ a]zw2zyi­/>|6U_.zP}S(!}ǟH[/.Ig2-ti6$]uPKF:-tuTt՝1M{Lk#ryQG:CH Z?Nۛr34 ^h N9awIOaIy*vJN9:?-LO8,Ǥf%P#OGf֙>8y$*_~ Oΐ `ud;c 쫙q+=.''3?ͻꉻ#]<[?oh;_YfWtwX8GB_ӈkLajrvi a;jod `un\eoI< RuQqQDB3XtI1HJIkf#_KsNpy0Qp+M9uI@Tt%5j#>IО^A ӓ0ttNY-ѪЕ$Z%3#8I) uj̞`$-{O}C4 !ʠ4QE&0A3%G­%̠%ao2ZAm= z3`hoC<#JA&kd-"GLkki @l]_\5lݣ7t , ׁ)x]6ԙ\^"݇%M/>w֋AJJwfi5,hbcn=j_.hAvn M5*7P ۰(ESC͎w:Qi=|*jNzS"I{g:vHY32 Y5Rq t)"\d!3e&r}kmW&dx8{OHd:qұ;3ӋKʼn#ǖMY.NTbX !,\V)/Ȏo -c. KTSZ05EP2˝uSƵ̠('nf2YB 3  C: 6L֞,rʖ)Ӟ f)-2ASR\Hg}¢qP 0#1تk ySIi$5(լNFOi & Y.Evh~9 (U4{4D9N@ Axͥ;3;HX%8ʀtc˻/!N/!>cw-FMuf5mARcs<IJHWzyn$HzٿW}4>8>`Mӄ5M P[;>5 <(ТqHGkGp%xCO2x/NpaeTCI\qX.x`tRKri ώ*7ezk6${ԚrŃ|޽wJ*rI0/MBZ)؜QC$Zgi 2&ʉ2J5qN`]{vȎ :.N(:%Lsʇ e:VskR+&Nm蹟W_LNRl|ǿw_bJg+[>YtLkٿI昅Ѯ.}ְ\2oV~ofWTɿ/t3 9j:eN &[6}HoC:}zǐ֟~eXxݼ>CZ} I E+"1nAG|s"p5a)|Y6, 7;X3=r':;/qPLuj ES{!& l' ,g BVEe\5 (Ms g*5<-w97A,λ=gmVGf*xQc,8-?^@2Z2kHK j23II^ #fNGC7STS0(ՆhZAYv%/|ݚ]}ne7]RڼTRZ4qEhÇނåTRo-:|Xa4y!?>WGY=|!tB5q\I٬zJ{T !OUM÷jKjdOKzRk_ꆷ}뗋+^i4;{#T2OSzmz.)V @JkrfՊYMȭ ק '^Fm$yn=]l.S*|/~|H^\1m[ՂIAUj!;jr-4U4Ѐ"ЭdCUdioVeTD`^ Acԉ"n.{~8 x9]^DF-rHyhMEn,`_kHD稘ҚT:V+V ۖL#eDo'!Q!/3{zHB/vBms7mp6v/#JuaEyh a[9R 9N*'{!y2@jϻq6$ 1@"1 b!ԝNNT,@\gEGy.J*t(nEHfy-ZJ-if^5Qo,|ZjEh&Z_UUDSO֗v6VV3PV,VO?/j1DГ(In1Q_H#On*Mmsh7PC_?|_$R* ,Iph""g}،RgF3TsM`NYo#}{]r8IjK9|hN C6$""o)k./ަPzy>wB]tH~^fȣX?6A2_fק7+xR$J/2jD]iEtAi]d QǺ:C=9@drvAU8[eJK9) J8{qaY/oI*,ik0ک)Y/̧4o68^:w?RHp&AKuf'̀sT|_gԽ9;:̱DPtbbE]|3( VNa27htdBK9-nW݋g ` WM'c Tc~O9!Q2P^}|5%^r"ĩ*M*ׇV"ˊVS>f7nNofjZNJY8˩FӅIHH&f\qU . ТZa?%㱲C֚  /(RL4+B&#Or"ݓKTAgqY:!B{dGf? [QcZqO^Z?D3*+"yGzwԾ|^Jz<憋bξlliO^2৿R+p W~w+N;S;>i 8[guNj?yM-|mѫW5ryRJاKگ"4^ %|6OZ|2TBW1|*, Xp!x "A _<2.`n Jy!#pCY- ]2%E ;rFh)[B9 91q|}}q'~6kPO99 b|n{sųV$v85;Q2TKETڨ&LJ\'n/mlEAG}؜(Hw<5\$EhDk"tΊB0CIRFcpW4#k|IxP&49 kU]Dz4C޶'=჋&*)b)]^`5ɯ !Ex!i&f&t7$wxLr>KnFD!]]OHq`us: PtJnxi7lF(tgGX\̗|)qo"70+>T9a ט3ر;^ѦLZܝ9J@mg zp'te ;ӭMÇLJߗn><7NC%*$ܽ"'x$ݧNj˧&|ye-^1!| lV UOɇ 1D{DL Q^Dꡨ)PNdo+M^|jm!PD4C H<6sA({+((É!E]P ~ NԤڦHg *ǔf%*SYRߣX>9XP;\2~IC-*K|e5aZ-oFIs: %H؞m|12 OQ~KP$ 6G=P6ҸZ>>5_(:ZtΤyQiKďBO&dƯj4lhgrxφ!/sWo3ƬC5?Q 6l ^bU1"/>4Y D) |D)5 ( ._O,T7tal"vJ+ozl6q8,| ٥Q*|/6+0J#4J-1\Ĉn.'%ܙAILA9pB$)kcQ`BJgY“*ۍhUuHiX]g@j+c+#0G.RZ{<;W\9< 2l!SB>ڋ7gL{D/ 79YN:"6)pyq&aJblaq7 dCJ]yLmxO64 ʃH׆7yW#\+@bHAb#Kybs n+4Aذ#AsU_eR;׬}`Kv>v% #S$&eyB5 fLhTjmATowPMɐC-l)GTdZOiFZxqHF˶}!TޤA58DpWjUMP4FH_)/&C/;DI\bppa~\g'׏O-6y-(0B\?"2)h%}eGTU_dחv.Nj"7ٲr)uj;df5QR\`ݧS?|*KxZ@01Q@Ve"ddH4R.S*T(δeih3Y$Ӓ͊:{0)`4X˔Ts*eZ(B /xњCvIo# K8Hy409%XJ̷}r O` "5;tI6.:H8>kQčmW'A3;?I8md Cd& "z]C̿ ~{32y {3'q0=b~O{ߠAY׻JM_N.<}YX8B`;|Mjpܤ7ծ;@9.%2 JAhઽޙ8<<I H(=7o'q_@ `:=N &a8|>~?|o38qՂׂ ._;XY !.7L4ohN<&*˿4URv0Bw[>hwvȦq%}2 -3Z 1 N=UI6}nA|K{Y |=IN ݇9T`4e['| :Cp)898h|}|/W)0]rGgp`!PmCy#/{ID 'gF"*J+*k*޽˖5XGk O{x,26t_ǙK'铥=5wYGE~^ߟ&!o|T2w uo?|f3oGiV` x?lozW`r*[~5|zmg/@wpFkYxgʁr αlGHn&gO5]fSd^,tVLy?w!޶K%֛ڮbsWA>,pWT%[_x3 Y8N gFG*`B8$@,L-`s4'&IRqME՜}mej_#+i L?L,Ml}W'Ru˵Ο -(;܆á yYn8ֆ|lv5Ҕ஺NJ;ZPZ_j+5ChOH@نGHqԆFg`#&:R8T䂯-@Ҥ=~$H˼r8fE4Cs^D74FAFYV"0&4%RpǼ6cYЊb1+W 5pkVagt0")W0(kihMsSRi!3;#opSNqD"FCY_ _-y|(@ efAb;76j L@Šp)|uCm7z"CHPE*!RCTn k2nA,:B|ʢL2"6Qd Z-@E[I$,L RaHוZ+i4ph] luG')O+6qwj _T&*2Ρ0$c*<i",#H*JJmZ%NCS2h{2it o@c΋^+Xﳱ?E"6|7-=Ddq2,InX6~H ʮ/&nީ,G׃ϾO!nxA.fϫ?77n^J4yM1<Лts?p:Mt>IvSi+sV'][~}+ۗM/87_[O|!xfUN׎6){7Q;wkA4}Fv YYڻ5~\ֆmS2DI7Gk+/r+7G;ŔbK2q{)t&IlnlĪ`%RcdÎC>B#5 %̙Rr7}P~M Dɏ51 ( 4K*btYIPslj07ٽuc96 Fa#I3`>N݅)Q%3m- ku HS:1MGKL?6,Ös1g5uZkȩL%V5n浵i9ۻ`~|'V93ګţTUZLQb&<bRY#Tr u=lj ZF5܀3 RHe;!ն~u9"60zIוWWεk͸ _vkC^he vna;xzf %I L_1&}JYI@fT٪z1g 5R!R<]*NyOegHK"u *_c;UQ;X2{rg-M7H}m=qt7m0V+/{vvn_'r#sLDXD4\<Z`#bC z*BrUPVlӥEO27-̹B&ѧZS-軪9y*Aj5+\~XNJP,cwLYh^7 q[0K".kiG \GD"@rUfPW2,$i:-F ,PChqeZ=2Dʴw(hJVL*)PX`-bWp7-Ҧ&TiΜ]Jڂϙ d(B*m5\v8r5;VOsM{QF(⍂B1(3&;/.`YE/? GKVrDFQER}\_(8kQB[Hi1bQ=bhB[OpIʸNB L.6. 7D !v7MfHP(&|v6 CAǏ{3B⽦4(TMۮT2VPCI\*9#(R :n697}Ma1Rg{2:Zԛo7|; szph8mwj[A~ !$ 8\l5ڸ5Pe@\e{9`YF)0t-J#:E~ GڝrCS+FTse%ک6Jc??WV zc-r5~2Fv&Q ,yK8i:nh{1/2-XBE`s;Q&!/|L _@vF% 8F!bEwGC'7"˒x2,`N\Bn~{Yԋ: =9 ]".q, ]f!ס6^=y匕+eR³Z'DNfCh_61"hIT1ʣjK#Vsg/vJ]e 0\lg0g1:L65e}&B,Ј{̅@%[ R"L˓W#7 bay4ψp+/0$Ӝ2߂M,!l%$WH@r=)(Q6|3;4G xY'-/tHT09 Nya796%S5mQ?X/ӧ;^&j)0^MKCj 3¿L LúdOngÄ%uP 꾗3+ݽ"J~I#Jk3'~|>t}-E(g;E&cL&n]ʂ=yISL nc882}sf*%(g 6A?DmFu6t62f0X׵{V&@\0W/PGFjF9ثBBvLb%s06@w1 -ZWeZ@E'Lg?jzmحEn+[>*XSy~dN:mdT~脍QY97,ƛfd˘'z@."$ R:>" s bsH#m  $.w !/NQ$c;{b|,뽩*_{S;5y]i+f0.WO~gkm٪`GAlpKb^{6a~dAԪOj.j?qCO7\}ńڿ&l碑ڍqȣGѣk|:Qc[_~T֬_j=5{}9=}U{HٛO_&-~st>vOܶӫ\yQ|u#.mQ"Z DRW<--l1IZy-m1uqF?7bh1(-WR/6v(v"/o/޾oޝ|&z޼쎋ⲹ=Ww͏{1Q{~lޏQctߛܮ|Ƈ~=7F~sst'Ia?.ۏv_J]_?Gݍ ?t=v+@Xzv1;nNèL/7r]5n|5'9f 9#Fdp(}Gst.3⫓C¦.j1{v;ɷ;ީbj>LtףvH}-U{5֔ƿf.8mկ驉љjuӻ_}Q{ pUWOZ7  D$GM ?lҗ~Q?~ϛNmRy_ ޣ;aW=8{;o4nUf=FnQ?_w'5tAmqZڍJyx3!0:ST←% tZ9tpQo#eA5LHPV?Sj=h9sf.XU^E uu]e~9R?s˜QM]U&}oq1!L';B {8~02D6}HUv SqQ bʮ-![@[9u%DGIHEV&WŤoԋ2-Uӏ*6&z56xRlMtDD$A& E(Eе+ 3s9d?l"@같Wq-+;-GeJsSzPJ0,1р҄ŔNT4$P7zEUdhknKD*tjt.w렻t6 SC-$:ה][6 , k9(."{q]=0 Y::D:l4&Q IjSW_-nQ]IJZ'D֥T.|CXAcrX“#a"MNSbTZ&!D!=WAQi) L <^ELN'4)3UIl(۞mw,죫1^r"A `xluI-_j/E(=y۶j,$@n,pq҉\,*:'(EVi,t/qe9, rϖPjYRLL+/-dZۦ:u-U5-UylQ@)[ws2L!uZ~/{~r".( o&[2I*`43' d GfXO_k7.e CAԥ}(}}yFRkҸ&Q87t~SGuKCb bqC'V/H&ti@l uޤ {±L hP>tLf̭i1l9:8MɤO_ߝkIP֛{ɤO| "\,E9wq""Z""6ζON&G`*c  =&oY3oͷr*kۀg5מ^LBXV`6m?=,fJMJ5a mtd[J6Ey3j];x'x|k)I}h ֚JZ?;EWS Bs))T%߹˙I6̋p"~[N@DђO׾gͭ41m)#ƃ"> ,sJc.þ,8-lLMZ/ٖ<6Mp|lwgwdD˒: Hpym6ei$1@lR6#eEPl4$dA<~1X$XQ_&dU(W6i$ph!Wط`iyQGUK\{mR"rk[g dxW(DGZ&p@*8}9` b+:/XAef(3y36j<w rEDxXr#D&&{ӬvA;l[~fe DZBr-p \A9' t0Pr1QJ:@B`[P0lQOFdA/ -ϱtrɥno&vuћ] 3$LRUWfRߺ ]Lʑ|iZ Ӳ✗^xق E~6s 00TMMf?kı.drSJ/2+>I`CDeq[] Ld=}RHψ4ADOO=mZ!֒:&i2 NaT<# 9FȒ:dHUWVRD.I YS$͡|3$5! 9Z%Fw/ݓj.@| m&ITDCMI}!} "^!DۉJ>@%aޒṀȥv+6]%0!1@C%`8@P;ἔiu c2yxUV*3*5\5N(kvC$C04P*KR*y *ęЋFwn T N[ TAE$<0"ݾ㽑XK:' 0Ո73{!@{q@$M>J86N:.:t F`Hfh5.O;M.r~iUD[\~D~սU(!ͭ.:uMZ]os 5/ ;.rTb+٭Nӭ"*N(Nj\)n'Jj#%[v0jn'B4]:ͬr4⺔v n ~_%2%z0iÅu  &52!P&e5C4ZA#I4j+5nӼh aH:<iYL([hjMSű#9cF1S撕*>b%r4mW$@wkHX͆#iC*4`IaX!,)M"c|BtX\-K&- )kBXL{+btF'Rh'B*7nC$W$:uXpV2qEm~7߈% eiA^: ?GvS.[*oW{vON(7 !]AMz3UsXSSjd[`rp +{YQ(X́(fegF—j-drR qF`1AF`ci⚃$|B;pD5^C;ԂO FB@~q=|eށ%2]c8ÐZxdA(&'G FP5q's3@w@tWފ4:y'G_5Ols+U@E0;_7z- &L,{Cn(?pnlE ksݶOg?,扈|Lqz/ςŶࢳ^&,0&On(a ȭpg0$MP㓏 A^!!o:=Iۣ<\6Ft~ Co03o&w*hgM)Ǘ*>{sιWY/y46nMَ~h\mS(188vfàotZҁ,ٿrW9? F:敛LZceȯv?wiK:tѡ)z(Z`R.fu25(gkqC|6ʂlI^ttl~|E Ɯ#A#F~ *e-5*mvn-̲$/Yyؾ}j TrTRtwR,^k*dXV!ՠdDB% ͉rh"M$Ly hwÑ(ِ9Jq4{u?hHj;c5֚R?pvNȎp48`y~'H1AУ#&h>DD6C~:'ud6Bs,HNEϵ@ r}ʲT[51B {Ni3on<6 'OjCz =ԇ1EP,K`xeYs uUPUeBTh+p$#ryy,W?i_BxuG W"JK .:&O rt "?Y;F7] a`8&'rU^cIOY(E'D3L2Q(Njң[{/m>T;tjZ,Ev]+a{0oRC9JC]5ٚ"姿ISil.Bѡ׶SZ\QҖzcޅw]JtLDbёc0%fL6$̣%%(ƈ(ÄblфJB'jꭠ:ʹ1S$46vu135/Qc]inCB'( YLO&JxGX#3`01'@ۑ5|!u}t{uR[A)gm={/j0*@H5ߌv0t^.8 w=&0Y^&TJ+'G7鿊ERu&/jT7W顉ɫf&_ĝC#Z7,ʴ%  Xz?.7J LY&oOHVz&!} R_eSAmL7 i+$+]iO@ gU_Y.ɬ;IG T 84FbUD$rG6LY(QZKO|pgYar+N@gL:1z`]O*̓-i>foޕSaIw`l:5Iţ4ؑV:,,Xx.A'}`g2# VԹbanm  !1ȚMD8DTc Wcb,\sz,\sYX+Ij̇u`t$BK=O>2 [nQB'Iu^4Q魢Ln##9>$Q*0e(9=])@˴f9ٻaRE ocDk10@F'cE%RN:0`d!:=X/a1pݚ:mt єn:U 0nmEbDbDb87**+Dv&pFX7R ~H7aQHAwYY+Y^r]*h#o rl0ք ]:*dC#DDh4&9c/XZhI6EloxP׽') %Qo`0o//ՠ^ĒbvͰz8I _鏞PK {rN ʡ-؍RJr8)'7&͋۰x-woYd0/]v='u=_Ã$5Ԏ{@g :qY/=Bh[ qK<XE"űlGUp8+!2:$eH Rv_YSGܲqsO9 r<(xQSSkuG*6jްdXmVa7N0JWR[V (.8`'U :F9P<+F3VyxK0N E"NHIysFA%ڑN TJ'HQΒъAcU |vavA -4`jY0=P9&vC8122a%"RoAzRUk+-޹XE8N|9' Bqߠdnl0^:NS,x㭎Sځctp_M>D\^cR>E%A⸝+&L%[.B u+֟Vٻ6Wr߫>,F$1=/궢8&%qHΈ= v [35_uWWUw}ٔB4N8EKcF%M,fIPz`*-8jQ\ĨJ|XH*]&(spDeWٱu z-:kRFeo>RDaKBH3)rk-xJqK[\h(ѡtT#1`l-C Fԧ "@bg*R8oNv cV;+sjaރ\ۯr;H68bfLc/!rcτl t؃S X>/7 ᄎȵ8ևc'Dϊ4qs9'+,Rޒɓ} gыxD]< *N3NWA0bggw{kVwK>'@>V(%0Bckz \LD;sP ^m56^b%(%ì$B: g:f߃ ;f'7~,9ɩЛ}o.=g ^]Z̼ y#"Od2 GOu 2[|RV{"1InEts$:c> J]NVUbWm}7wbŲ7)M;޼UčꙥτnfW3杴J*HK&näF:#1(VkN$h+0^.fp=ƉSw`/bDU^rAqI&|\2 ӿ=.)υ^rЖ%eǚ@LĠC!@JУaZQ %VJ@f഍@*4Yhvș\ci5hC "+-DqJ:vDEƺ,Q |H*`(6qj$VSn^1;!Q1:Bݗh#}#pڔ N S1 %) %%NZ9ٲQ1Bcc*! XH87T8ctUD;F+taE)"jؕ x$DjZUpD0z)Wd +Zxz¹lFYk@Uvq)xB &;,ozjtxɄhPZ4Va6ggrh@:O%.Vϙ̸M[=83m6JxיЗAmi[ -5Pj!8168TqiflF(EC9<<fˤnO#URa8%0AxOBZě8 1E1`jLe (uxm3 *cZ("MbqA[1R?m'37˃Zvrj_7R*X|֞N3-MQ!k=xq)pIrҘ 17{Wfa~Wꖓ g /Q^Pp3 9P9K.Vrbk {4Ax͟M6J&ep~oۏ}|˯կ^ͿWvQ0VOQ_Wz1^3*=kT5zӨGWӃ_RE{Ր>rJUszatJgoݭ9[[ܣGoBMXw>>}h80qAQO >2M h9˅$bHeȧ;854eR ;O撃&h(=5Axޜ2O^",>~>_/bv]^Ul&`/%M'ZYK7Fxj7;/HaGBSۯ+ ^#xu?=ާ-O sI*c|}g%㯟"b~q>`j77wߪRhfkO}00 psZ. C)sEHmݧ螛a1]eF+h">]}Al4:SBJn )1N5vӕ6!$L39`%'I}f*5%e&6.,ijA7*RG3In:^ڥԱyN4U 4V)K'-=|aסxzQD5TsŤoCkrS𑚴.eLww&Ƞ ^!,5aݥ9%ս (zk_%Q٩ӕuPqcJ]ᑛ|> 5Ji;\ݢ՛4{ RR5po B&|*.Ǥbq]-N zV11Aɸ9u|̵n85u̕ЌO)ܜ3A8C1BJ$˃Е}#S 9оc,ə@] *y  )4tZYzςJ Z~em;1wmϯ⺒iU>_U(A}q 3#sMQ,YdvI r&h:7 TJTn]7u7ݼ{劆ٔb,Hk"H5k,E}V M߬~`V˜sEksH-z3y=XrɩM\:lml4Gw 8/r§#siJg\ %tkKC38:\ Nb<59g#`FkzKFL`FYD8JK_-A%k[v_^;A!]B/"%7 ]vyG5hi߄+Եd-;{ʓJMDԌ bˉMKfۍN[1 X1ߝ[O?.?Pv3 90W_. ϘWQ#pJnZH?OԖtJJY]/|?l _?_kHWP7=,;Kw7Wg7.o޾=f?wvȧ%0QԛOv , ͠2>nΧC9|PTܠh*=6L4(^Bm{*-ӷįtʸljwmIr썮ڀ?9dE>n[Yq$9f~Ք,SP̐qlBrXOuuWUp>PW9`\P^{-K:t1 [7(nDlӁyn:d@dx(7X>$y= b-S/߼d5kL"Gf-eVa6uT腵 4R)f6f4矯9f, jһDIey $yJ )&WYܯ)ˇ%^D#K0uJӦZ3` 縊a0 0H#0gg<%s5s^3 6O%srϏU< =0gSz\gﳜF 9ϰBZ~ CG;2|Od9h#I X5⹍2Z8;LFwjЇ|JNBSIJa*QS2"w EZ!h0#c.Yy`ZƋ(]>ҙ>NY?,ZzUu8w|1p ^,P>=0!zxI P -:ǽ(bH4yA[EֶQЪ^1W@ƫ#("yIMm5.$2ckыwd'̻IM $* Y$H!R$+ADŽ:Y N$ FTؠOJ,u򺖆ZAFpx'Ԛ4FGcOo[ק{E!꧛ק]8هbf(^e6(79vk4ՇKG@[CVo!Ѝz#qA(j'fzgwCbbw 7uf:G`P1#2!6ƜsjlY7 |5y.4幣Qc:G*L I.'n}`M">;J2,mOV鑯́fWT4 7ܝ13Ơ=٪r (5=?Vfm8T V o,Tr+MސmZ&mkaknqFKG= 1A @3  %)S 5kӴOVW+RY?$t¢v[³1^Jng;NӶ<<ݲjMZ*`] eݘ&_߮MTj}%PJy~x!R?@~}sPٚ5^6o}5kL_!:7.e>DylZ> y KR~B]5trd՛)nMӖmWݷ>}II|3z(Q8≃jϤo FμF@=_`yu~{]wݬw4$S}uQvAS9e^+߾n:.^C5?||}*Ad3 :I/ۧˋwOUQ-~N^-~?sY}g~?]]3-I Y_?}x H̯mx}CPMp} Y:`jDp{٧c_WX`vT0}vƨ}QjfHX% :JǕ[!vU8M!6Dv@͍8S G0n#ѓFr-=uޘ`}uc?'QJVn}{ҝ9 ?0;8C4 9 aTg58:'~@dg fx(̌ԇV2{PYLovmCzWV2bwf tJpeHʩ<owfPr;;,mʎ|`8arU,@{ewL"Zˊ,Jc7oks՚5V`l1K,*ynb؝|ڝ+>'poYL H {-|۞@#-wy{;ve蔣aHc)6x@WwI0ڏZ>Ͽ}/߼:ƫ/epqa_9Ӧڛ䕨Vg\RB9CU%Ư_:$ T002zn% I@QA+<ԧl5QGm:SxkgۇiWN47JV?[h~: XW I 5V3N#P\fΤCk9S<a+e`QMïm{C !).-E{E*ݼ>KO jֻv_Ky.?|~YA$M1ec:Yn%W;aEhQ ǀlp0) 꾋o-N KbOr;P0zB8߰{PGXH-%h#q9;ayΗøe!4]D عR' PGqH\3!OeuD߅z鈾o-B*tpi)aIVɅs?reIeR>Op#+Rr`%yzK<]YbIP/I֒j*ɋ\EXdTr?rK.}P翾y] GS52߬aV 0`x0H%qИĖ>IS4pz&^,pjγqe 7ٝL뜐SbJn lRHȺ;dӰOi)?V}Ե%4>sḷ1$)4U$sSy݈,PEEU7;ph(\[ bHXؔAF]+xjϡB1{}y*!%Vެ 77.er:gQ: iTs4]Mzbբ=1Z>b;U>~szu`/4Yx}ځ[lAĊ/C)A֞wv$)YljXhL86:D(Xs, M}6xZmnYg/5;ƧYçk;]=SUx鄱#-{%'EܫF i-9²HB"2n8ؖƁi] -=h=hGmrA5ݥ7.JM7ZZ,wg i%֞&gQu;K\\}Agz#T(7;#ԤFR AFg ヤ1$#lZc#Br,K;.4+ܤ,c֞ (ĺw8uL)OR: Q) qe .:L &F!A^Uz^S_YT!4nҖBLZ]i:s]7FK|l{(BhCxqi[#F`&ɚU70 BmS KEF )wWԇ$(2b]2 f@F*$Id6ц0uBbNOXYѹGL(khѨ\"7u A!ImS`# b/p8RΊe̾mOs҇} uV oZThP,vf[ lFk &D֌`ۃ +avRbnwWBk Lt3hE[`r(Z*7 8e@:]SKNJbg;0$liqF076fD<\`Kgłm?²nK#w u.NVVr-P,uxgU-R&:IȺ [X'\r+%AFଂRwݶgj:!uVY5fwyo~v&]H )` [vИu{_E+oXMҶli6 dakm9EK"#8,`3{y.fwY[Hg&AO5%ԅ)^Dy;E5[WUSu l)C[+H,Kx 1uE[ R7[J1+2 LV[L:D[J4q;(UR͍0 aLG32U(#>Lzn@NfX_Oe!h_b׻nx߬uдa4l>]5EUH!'E*6zMQ$UVĠ!!˜ri#=yBxl0+n r@ Q%{J|75usN`k˽KF@>-r\#xY0n* v{YlmV88ƞ;"! Py&U̝# 1T$wjqBO>B~WJ$AX1g#X1$N3^e |LBlSH[:u0jX3~"E6qEkc SA$}!=n%RsMgBkı﹜_\ аgZk"!x^%Nw?LJh2=)=K j%|&JfpJub٤3Z IN_^m_9J C&n*pA*7U2ʇr: \嚺/,~LZ#,No AuM eMǭ*'B}:q"Uv#N "N_*x 7߇{IؖA8^?M toǓƣ['.G/lJ(7Qd8IeSb>8ǏG÷~;=Ld.szLo㱙K81]u̮3ڿ깡CocT-[ztx3 `;SNwN&-4y!I,&d6܀{~0ajYLf"P48gڏCMn{U0A g*6Jge8dWS*S`WfÜդWRXky4Y%\שkM;X ?26IC3`!M%PTZ\OTٴ:jxFtb5Uw>X;%S'H\eq:K_O3*x+ygf `rԞ'K=l()=EȤ{|3\ /wg!Wo"6]V]׶ul6Zgr "Eu/̝Ɩ`G^Ʒb b2l~34z޽yĜyqBQ#'qNOFQ;L Ez`CQ4E$"bpo"},q6.93Caҡl@_ %I"qZ(D@P_Va$> !E(r&~Yݰa-(c { lw%EU|4uÓhlݮ~=ݒ<$σ<)B:Iqhrc!͒W2zZ+ra5mRXM_:) ǠAyKUc1wx9|P%;RJ=UC?6f !7th4ȍFrƇ,xYRZ:-/u; vC)H'ֺY8EX5AHiyqo8fpO}:t߁V[AGcĀ 8*ˆr>/EFKeS0 s'{c(n8=2Pc$KxW1bdz.y¹ $?^g %{jDڻҶڝQp N}\@Ύխu"SE8BR&9d8N|/!f9D7笜 jy:i)8!ZM[mCL05ېgl֎\p(Um1P׃V$ϥckE%oK6-϶Nm8 -~ـ[gNz8pLFMx{k¹=7h&RYKST0k}zΨ"^g=z+VHVmG ͅ΋otiq̮7D9>!~Bs= 'Л{Aۼ\a!W/n(깬G3K\ k,R[/4 R |( ՞j#Xv-,9|OGFb A'ͨ4cp5]kJq0a;Sڢn^;QBlV۪l=^PkݞgLhzHY7sL)Śgژ3`u{kZ-n^{ZY_f ՌVt JTW^!TF>6zF-\`Mѫ^]t@PQW ^kKHQ^UEb:\h  kfɄ*.廤ًVN9֝}U?yû=iPgm$a? )x(KB2Oy< NW}'L’xk91$Hsµ Nc-fZ:G}Ot ,$߇ww+W>2EbB_!&.27q }%Us"EB'N8q81eŜ K?_t]UYdLYaJp ^'8XtOo{^w\eJav6{p9ԯj~: '>UCTc%;dB 았D$q Rύ`ÇZpxz#tBZQԋgdhSSk2I"%#B3Ip'm -I9<8(Q6fTꊕuF?ݯ)BJ0a #2IXLOW56C<(~Jbe%@4l7SA0BZʈ7/Wu{hHN`K E 9 ÝC +(zcjn u_@A}I [v}O~=,UY%7S5]BiC0lc@CP 'XQKN$2 Q+ߋ en<> -#u&}'0 0?GpO/ޯR?<\H驖#><'#M1Ak!-VG~|%tp+?nwM"0Exv JKP]wf:p4 V-֜x*6{G8,lxAppH"*/lfƒ>ûeP0^xTe@9&,%w߿h<< ӕdu+x~>4+.I=,x"h(w $69ӕza*6k.pE^SQt}|T3@ !1Bx 8g.EJ ·ֱQ1h[Li@Ԭc`a \Vг0IP2K48RZ%@ɘ&~\g$\{@Ac^vIKLirɩ7#pz3ûM0\-,  (y #Lbz7~k]0fDsKAhI6-% ےl1㈀k)a Q­܁¤K$TJʃ&vV:ŕg R1`j1 n20^X1 h=NFOO٭_wfc/=%b .ifif=ɲ#1v'ĥ1*QDi”L={X{fȂX5ٶ*m՗̶lmv=)Q&69xQ{n8y-{ ṧPBZZ|J/^!slqYZV`L:KKc& i I׭5 -^l eeDؕ C%%B,fc'+'ѣ^UâzP…Cv%v/%YyR8_~ go޵q,"e7Pŀ"oHN 鋭 E*${$E!Þ--SÙUοAY(z*b+HCmoF }.،7]Q*!%<#bT>ی@&`)=l^H`ipX#bNy #(Ja*uXh= ۘD aX>H'!EwW-SV``*Y1GSUTH9,QP5-d"F~01 [)Kʗ&,Np0aѠp&0Pb[w pa,wj<d&8Zר@x0-@z!>P?Xngz}.|tJ PʂgVpC <9.0 jzJycPkgq!bVZ:^p+B71R848V%)_+tw`F3@`**%vf̆1C~@Xy`:ZsAHdui%jVs0UFy_KVm#hՖz @}s n19G#8GX` YJpbYsFtU΄2kHEv .pr3[mwg7k$WH#Õa!cUj<+PAAY-Tg+Jy1` p*A# >n-2Y(`,YrE%Eb_ަ6`ZNYm#yĝtM6%RU.ϗ 9v0,16`u@;؜-(l`.ʓ^yHq/a fI$vl|9rWaF!$1e_] jg8a7f؎Ӎ/b>h8[ $bxgY6>OثaG?'c]UE/c#1u|$.圔q8{s{ǏG?[S41`_kū__Ϗu^5`࿷z]x~?g d[/^iŻK<'_+D:ⷕ-sqҒh#; it>t܋I_󦻾^}3HÀ_ðsxK xp4I#*Q=_I@oaB)(FiQ(Y^Xpl{r&l8\IHUeܭ[e{;:`Dة 1 }t=pTT>#oZ*鶭o.6^bkߕ g}nӵz>sSpռ+>ԕ}I]QigX# s!c"P<,!K) jsѾ8p Մ/\W4[$m3OUĥZO^bX*O5oEI%qԊ*(ݷJY!~2%A$w<ϐSŻK֓՝Aڤe ]ψ8=ؗ>֏>0C\%5NX K~$'?Wwj2O~jZsݫo H?~7ԫ$N{GuZED4n?*u\Q7Rg:cTF2m|mekPrxXxT&K5OF L1NDAe3zaRʹK[IQBA<…dDC(s&)SZx wH(&"sqQPlO{Zhy$v,8\ R`"Þ`x pA 5VxT gLT&T橅gp,8ߏ?XS $b | 44G*N801cms‡u1'/ިA0fڈ02PR 'n, J\)Xx ]0−)U0ʬ$ƇKUK uqRPL7(Nً0x:s`*ch di6J]rϮm 7S/ N+N :-BTFqp:F1Y)"` QklbѤsZ' ùCbؘf"[k֜7A m6-]#ED1Z݃s'@h%)-8 ,h\j(VS ~=^jƊh=ܲZop+ RǙ %;dDTQAؐf`͐-2$He0x?w&U4l^T3 OYֲ`Fs7ȒY_޶%wohR4{32;h%}ć񯙏ᓍ"͎ɁXm۽_n~OwG`ǝ *"I ZZ,310屦8֜]*T מ~EӮ,FRT>-r8K {;\ Ǒ_ +5{, h/]j'BZV1dvȘHt O>@9š Eq)2Z()浚4Rj/ujI{#en1S,^ȘftDaH" ]Q~:YL1D(yUEw~J59rhTJ`і|:p<i 4ҬܩUH"X&kf8qKrF #.mbnJ.P4uOenH(D08pp()e,(C/"j汝R*}T3V2!W#Kq~~H.P_ ڝ]wɛɃ?'7Wg>J-&ژ_1g;vH5 @$/~zSRy*G󪉅~kq2u4g;a>ED(=5+ I ,RY-i)otB햋A>v;]?pj[ehvkBB^fɔf/nbn$N3jiO)Aiڭ y"%STnr[.):i)FҶv~Tքp=ZV U2^g? MCW:]TqoCzk- )Ѐ 1;(xSr /nx܀ qDG$+:Ap/980dxVGG|r }o$};Wp:iWݽk}oIqhwۉ+I[}`n;9ſK7O,IlQ5zߞs5Rj՟.Jk>ӽ$Y񌾋.BV2kDY_uLo똾6^%UQ_e ˺VD;J ٪ͮ={<68daU)@ u..7 %P . H Fxh Kp I;}0y,/41m{Hβ5bÖ@%X/1=XJs 5Fxu[{+n-F;w9> bKn Ǣt9DCļr ͛짇eTo8B7K JP"nnShKjc dG+qv|/)#Ur0/8C%|4&X*b2EjVPJM%d -XTLR̰.ΔPM{y `dؿ/qܝhcph_KH a+?f+M=eP`3pOGi?r /G{rKl%W-WHk"GΉHﶝ4UttQ (B׫ nv @̧~|v9x&f_\};=7WFvW/s~6G2=:{NNU\9*h RQ,#!xքIS檯\'y+T>P q͍=c-jbBh*e:Ե`P{&C(vhb>6~gf_^]`' )&* T4BgQV1aM`}<&&"q9ƂWw=_ wxG?=Za,ߋnIϗOW}oy}2r'B06`Qp@Z*E$ŭUE^J; ї#^qtE()GNy(JKiEpsS)SMJG52kRCL`-Z$' %\t i͠ w՝V?¹W"~=LT)`I~'#SpFiZ,%NN'#ztvDT_,ƓZ^xi>F,C~s꣙~]oHW9 hX$v/A2svl#Qd߯-AfI,տꮪ#oVj0x5oyj|\ BFph` z0֔ S6s OqAKUDK] uQhIV)3 6rqWX'RJ L4n@pqIO9Eƽf>_KZ>:4 EVCAd%!{)&= ?A~hjqD[2EY GrYh` R^[B`hc3XqNH*b\MCtʈto0q>e3!w63\#X0ǿ{-]A o7Oϯ yaG?:#E I)?z}5Jdzy~<|{Oqzsoُx~W;6 -OHG?`"5 ثw;}q_f C5aNk-=KJDЦ^]W\:F]ь>: #xxZy'=h[(a'S L$\h%)xҢ-/[^͍a8Q0@` RJ)pⱊA+aGuKQ r h43J:Ъ_ύE),S`V5I\+WsiԋW֢_8s>>[S]E?bR9Vy1'`Co٩n`'7mYaݏKWDU_#da `e8+'.t1RtgX V=~"!! f9IT0-ě 9"^%>ǽVj1RݸZj]!CZc]p7XUy!c,xw5`>fxHct8uw&FDQK1dXy/滛'ޗm}̊6}Ѱump؍FOz_Vi\Vbk1Һe޹?R!-]c{=4(R`鍅]`BG|H,ܶ"JӰwP;4Ck+&g7h<7]侲<hZ,RFq#FED40$KM@4A),D Vf:^O}E - »u,1ߛLؤޑKm sj/ꟍ1h[fWM߱<UO<(|ϳl#Oڈz쓏R ݛzbšx;2g23ˍ=˩ CV`2'Mλҍ2Q5Ši:ڵtk(ژ!)zKbR/(Լ%(6 dwqOfk&iPz9'jRvaTIkئ&(K&a`k-W\M)t]rۇxߙ9ˑbP2uTKE1RoJ?y~dLRiy.NƂJCVKJmI{CdH1)7U# Q }ydS 9D\+pY4H=0m1aqQ c A:E/:$#R!- ( N*0tІΝkc11OLI:]PqgGIҿ+j[M6]A<(g{LSP"I*((ÃI(ӎxU{c0VjN$kS$mj2`xڻW,00sW,T"1 4:p< ރ` $ "i+ A^bJ[ =~;>Tk"w<]GrP;A:hX/4K 9 o  7Q1RTL-*NPyPB*x c0>Պ$K"Ib@I@8Fޙ0CQn`VH-"Ky qo]@pdt -_)[ԫCKp$=EnS]r.J6"ͱٵ -T2yu*p·;JqOfB$W Pӏe,~q0Pe<g`Vx>=ueP~ǐ|>z ? p'?|\©E׈seiaEWE3$=纨;=PjP74;hB@|M?P>}+rkv#HHɕsV׍sNM(bLB('QN7L8ꭕVz(Yk43y38cbv*D2"ıP0blz_kuR9! 8 :0Pl#b**%E49l7N׹f_| n㲙t@JL ٷ4^5,l=lKux6F͍c)Ɖ+2XEFWƏ./ GqbfJ(r|Qp.tv$,H#V{?*z5qg9#/^+uXG7W2>,\wx_LJ ̄/hFp32*1d???x1v,5-"CD-C"S0z x*k0#H xQ45Y|_|h'pBa 0+cV[IX3WJt]@Q!v(jPjض8%Nl'FqL_{Dr)*v!BxSwDbusiW֙uVt6}:I朥0+A9:;4&Q&084%ى]C/e& t=.lULs]cN^)7;;53cxrIa=;!Ɗ$u&Ҁ@6z a-HEaxHD Մ=ܥJQMHApE HJ8-C3슶aGq'@E\`~Y`LJ# AP+X#B1Xa̳)ui Naш,ur("2` =&!x!R ZPi,P?}]R-E1*P[2q RA!0(. Z#X0eQ XĘTQ2!m4#f+n`w:"7*|7YoɬfӻePҪ/^Z2 wO߿4<`Zo7Oϯ k+>i .? 3W`j_o)NBwcdäZ-N5?({t$gw]-RjFm,6h Xl$jVH123%j F98ќpQ5.p1H<6W5խjx%%+LvJTy:s qo?(MFSIɚƫP76Cӫ%-hU}20CM.,-Wk`df"N$BH PpJ{9uzL|Se#ϒŧ]/}@ٗ!D%㲘ra=HyEBĄ+. } *džjFS!~160T~/XbXŊiS}?C'9WZ'3.d1ЄgzCrXti 8H F;bA^" :a ׄ VQD!dgJ8"0R߇am) OITH*}9%p$r_UWWuב#C9,)-X/%+A ),(uE'R+nnDcB9yD3i}Yf25|Mb@*X8.,;&FH'2j5/x!3y!/AT=ѐJ E4r+^` ؍ ;%A+= TrCCL)DxK%  _fVDW5usZ2QT/ Be.)C\M,̉XHfLZ9Kb0fy6r-mj`&I5[PV\&Zѭ+Ԩ\ORSQP]ˆ$ [ʐĈV,џ!2K(׼y0xĤ|MՑ:x _~nפv-u( B㵾јJ;oqK@wWs'{2 g.dրXF$M m-m5u5Nu KJwu~&$R#y?w<2Z **U+OI +>ݣChVr^r/:_ ӡbLw_xso3C$WD .܄#0߹ZwiK?0/_ k x7X5yLuq)"GJ ;Ig B۲Nw]ef2JI8M/pwrD7JKmḉ*bS( K%7 Vm2Ǔ( x1>9L݇\9dNaWgB/o5Zhh;ΜM~Nˁ̚sϜLt7)9,p8Z$9;f;gd3N5sL:,5;{ǙK$Iz[oS&+dg靯r^d1 ?Ph_^B-6C=O6걂-w DK-:StؿJDCJev0+irb*-{0Lnt(C):-9G`9Y#)͘O9)J`œ2FUq&S4b )7$`o"SbK3 kW&,dRI.j?QЀԩ w<}^ۀZP'/H,^JR&CNҚJ1m;ndVLn-Yw,EAH}q?#4`ܢ1Yk}wneNww<[BjZ`jqJs/vZlRɺrdg]7, }uÆ4f3_:3^)k9xxǐ!'w(Gnd mA"ݡ^Ɩywq+0`$ڂEk*P*a W<3P;[]<%??1b~սy<gRC//YsaAgGň}Kﰹ ]7}?r=iNY^u_uy$k4svrj˥m |j. ߔme;y}\.GB kR\}LʍKO0)7h{R*^Q'դsK+AVx]$ 7GD\czJExJ7grh9ߴ SqHZևsb`˷OcHUV4[F+fcrN9X<$G`?$ԤSTsLDK;aoVúc߼ еTC 10oE͡܎S(?V2bss Eb8I B8J*\0C1Dn5Rm~ `hi58F+a\@Pσ3,Ssf ,\m\+5ȕS;nV".zӔdy;[L )#(8I P#FE/SR s_F`]c* fB&D _dH8N2dr*$‘3H@)K`pP-K<"-] Ú*76uIST#sHG5UL XP#8 Km3H&FrrOP-d;$&Rh?~āB{|N@PYt--O Z`$TzA5w7)WEj̞ 8$?xs&wG~H^]څ컴$^0u?, b?|t?Fۋ8_SS>=܏ôl0*˵nK>4>fh}L̛$5C.>}^SdTa-l$ϥg.Jl +ZZW.MdJ 4oC1\I =vkA4}Gv:&޴[j6$䕋2悵g%0߃䥶a݄,%7j%@7񟣲-K]mk/]%uU60Ģ"p(gY-GG9pW'07i't]uՕ'Vd%!{@6SXm2ʈP9|`?㺦̄Wz^dYnD:!1 E!ș`N QcQ&s`P*>c:/:XUi *E~@C"b\-c}FbzS$@U.OVkc1D"$1K]F%Ʋ(dnh{Z%+SE.~_Ɉxtڛ&(_S>SBw4{| F/ҥ.F˲m{Jk;~՘""+PA3)UVZx2B*aĢj'~[1;4&UM  xWͅv\ f.'9yWLoL|*t{.—uo?2w3xҎ{n`*4; 2KssQ(6D:9ثZ`=)œqc`Zcq\NIu]+% "Z=BH' XR *:38HPީ`85@󬇶~rD0ĝ\2a í!M NHV%0ТJ„L14omzZAM7)"Xdϋd&qFUWӔJ4wPuLJ!hnq/! \KR@N6ɴ-Wѡ&PQrFP52͠gUcm *ZT\WuU:*O<?Ĥ2-|12ppq3Vp=' IOU.on|/}}?E(--C?5laTn%^oqQ@}zϝξ[QP'іDTn &^&ǝ2?%[JR0 ):aq?%j&cE1F(|NRȹwf):̂w_ VU{3p!C.IʂyKAB+ʻf*-]b݉jZ]6>d|DG,td u_XE-ET\2uSmU~=YaDr'Yheqvսsq;m(T"S:N>;˞_JJlPzimƕsm͗n:[lZ,]Qspw-;E4֋! AiVJa*Bs4,U O\ X1ۧ8aOa„$BOLDJ+eY$" 9"1Fo9#)U^PoՄ3c9o7{/ׂ#+a4jB֘8h9WHnu4Q6J}iIHJ]0)b1&U_:`mƪi2XsD[V_vYUͮo!N1# +Y\,4= ZOV.Pvto<,B-%1>Hژ?YjaűugO5cƅpڕ;`)Rz[Q~Hg SxSʱ:`^slӉUu>4Rc5Փ)ヱo g4&;g.ۥTd,`TȊS^JucviG}ND./l#ߒ4\9YxoM,`65z<=23wH+#C&2E|wp'?_F~Y5ѥY\a3BaIq/^˽CFI[~Q3kZvM^c J*:կݩIīwU*Z:*Hꛢ%V*N܌[5)\ 9bY)זzCj$nA0B0m7oxکv%i,:. (7OR3I嗈 Eoo|xx̗"Ŏw /gmpXXN8N 6IpP؊R,J 9f1iHBַ Mxʢs_xQRxqdFWSXIh=9@X!E8DFٲVH 70 ߩfr-w1VVkJs[4w28^-bTo(YSt |09g !LN $ S!#Nk`v&-F_5a-4SDc$ڂEf k*P* uB3"TXaJ"iH(hFaN[u&#$sMb͐7,cXY%, U#0ukWb,ۛ Ug_(٘B:cU'L$޹hku>X?ЦF5{5T!B̩v0d[$bؾj !-KmBZh:/#аvZw{+Uk_(D5G\ݵ1Y*4AmrIjD$YB[A*&Cʍ4H,ݪ[{4@xG{ lT{e xw7 ?>zuih[^C'XZ! aecwHee*Hnm++ Y ee1F jj`e3Xie 8Uc+{wgJ+puMbwVzI6tǥ?`e67~6*Fdv/ZϠK!z5 <] iKOL_68MՏӴNӒGc4 07Y(x)!ӗƻ"ª'9kæyM*L6 h)փDHPpA`m 3rF'E6݊ l?8GaCw<0 |8-?) FY3Ң!!sPHBkEC/N%ō UC P@ f9)T0H) ine Zq@kPj%l/JeM.APo3JCG|>0;3ճzDzzm1)p:"٫SrjHhhKGA ]h`F5˜sٷ=Hͱ$Yf^z:G|=8#K%`! !\jvJVuoG{_v'@'§ʩuy+ڪIru9jvQ (&j6 X3U'Ͻߖ%5"7&`ξ!fJD+@,"{vx*@Oخr/eHBۮ|)$ww_ޘ[s;0z\aVpUբ9x|>q9j*+gI&@b*HT(^z`!P?M{i3>,wf4XxH+Rjȑy?Z0ig:^j8ƃ0Nt1M۫7paI(g -P"(`(Z٦%O0~i'VFA-%!{G~a}Z~DM͢'4!} Zpxu?صGsT7= |2#g!ıډwTRy,_ﺘ`]W&tݗo lug`,oXi֕~YZz~4/{Q'9UACn}%PK4,5Gs6/kZbP\;De -%Yv@ⰚS]Y 6P_Twzb_eW(_ŧ]UF}|R jk)K IE$ָcJv=}ۥ;z'XpaZ;ëV'P/Ǝu^+mz^^ʳ X1GmB{n1"QF%`+)+f*dL1|Z*|9c)g#@ثM֜"A<$!$ S(&#Zb/&t<434R2ݾ/v [ ^UewM\ .@ǝjQipbK4$|-dp0 V и=}R:D)o9q5ƵY+52GX QZāqdԵfVn#R!rRÂU@$0u*D+I6EE:JWvWlrq~X>٧ZAiƁ( BTp$Վ $A1iiRin l:3,.i vK-G=2pRavݦ ''WЅ5-_Oot)[yr4v,+7,לꪱDn@Aw0Vw3Ěޭ MtæTM?.Vkpv|c N; GiR9GAQBI DHE#oe/R{VnQ?T*W+d^ П5t7g3 ޥ~F}i\㷅 |o CzyH%sLSR+ܪp(0pBǜHA7B-`АC[i#"y~KN}1׸GݽFL;r asҏ}vOdJXwJZDpkȽe8H % 0MaBiu4xEʽ0cEiP0lAJo-gXAn9V}A\yn2Bqn*j8G)Q<C"! 'AD-< YC7]LFD%8tW]F/  ̕ Ue _0B7oFZ~oW?}N3x>_J}dKLeb=.qYe=>)iNe=PrLe-'D#$6Gbyi?`p}\|WL<׻pg߬B\uFpoy8|Z#gE0@2ơOVj$Kɔ EK^ ;4c3ũǟOC"0A &y,N!rŞ>tȴ+΋ %PWy}bƐ p(ʅ @{XSNZY<)I_"PCTX~?N r(̥i2M)Vn`,bq@Yú(BRGӛ|@$ !zkmH˞]}C`{":ɾ,E)Xo5IQÛÞJ [3_UWWUwU L Xس7b`[hf>LlVhC  m !F*gJ3Z҂'փ)jSFXȈ-Pj UJw~g$g|EXqD]3)kMy^X/b⃬\8BPݴW-xQX|ySbT 'cu .ID=׿&_vy!hN֍VYpQ~5nSTmzTM~e:1+@ PD:ּZ'A=E񊖕=|탊`&L+Zxaj͙h5WA'Ⴁ/N!^d2#_䬜ᮧo9X3rWBHbe YVxk>i}4# Ps(Z>D"C2 \Ob Q+9yWC>zZEooHwLH{z 0מkxj+;?tfh&~]O3Lya]⬫+1x&ءN.a,WfWb)E ZVT$#i]Ts0iN3CJ,G&5({ϐd %xg23,fDZ57 !@pvhLy1,%(0`rТ LEcHX:Z̰4FSBVEEZ !)NӇ"A$!Ur]1_ ؙ Tdߦ^܄l‡TPaYw,PZPa2'- $PȿDNA\S岵A^R퀁 V>unlQ-m)T^ /~B +eIrz$=bNiO=,|Db%qAr`.+9L0j(1\(T妥\N䜈 J08g1v@ĥoD#ð3G.A)B#Çpw!%OO)@T\(%`o3d{,G"Y"96qDo7_T:=>qt\݀=":'3?vʻӫ gwvdt);Kq7D.ڒ< H÷Wj\TT\,qC *pTSѦ$m=]{>tiByapPx*tD/8r']jP-`{o~0 {1VoϧN{Q< Dry\`Ts$)"};_$K8d="|r eլscf^]l/#b}Yw?yP0r4?y(}/Vთz>yh(-[<]L:́훊u2Ga`q䜌P+_F>B%H&[dCŘQfg#ea dB+OP/pP9e(#W"0@̜z+`j5.')%ՋX KS#?F1V&y fmF3;!,uFmSvVO%.+ai pV LY c|PF!z4N9\|4.pԺ9X_wȎDYE:Q|[;=F162HÎ)R%h ݛ="+MKaQZ yc3[a(5`0o))aaX Fͬ4@˳t/bDKDI%Qʩ˰!JQ!iو0r;lmLU4zqV7cNkT4; 1  S"aX1[Jk(@%o8[ + 6Nbʣ)9nȜ:TZ_€BfNX Q2$C P_\#A fV&@P3eX LoIqF`^`頋tr{{1/}w-$o`RCW\h/_6#"=!dM.d_온}ƤaB="\?Z&]eH#I1ϐ6o6)6׌vG_Ew88&!X|~9Y9[5)7 _xLA(hJyGtP.,,B#FXNó51OE#V1={粹$[|`YHbsGo$SFB9K+LziCHS,kNA> T'hɓm~s)j0:hh2Uf-n \%-3KZܢ:,d ecNvΙj-(\9;X%G\ Kѝ}usط[P:+N8Y5##u+K#φ uמGL:aSՙ+땾;FN8nr3*r  ?4 {_ջr;u<ˇ<<yBeX?w^egxSZ!p;!yFᥐ>v~_U*{tXD%R-)d=B¯vx425Q#<`¼{Kf 3 9Ҥj%3n,587½ XhGw9ݹM)8qx/<ѝ+d y7T{Nb䒐XxUyKf 3 ]֫ѝ3s7Ux=t DBh yE!Totz5 a88 E^Sy:UH$i\^:=/k.dUHsf&{Yqq[}Zk 씦Q HqaQ>~yeq~*3ZF f?ɺ١b880;f yXR BbQXPxk 2 {0wRac*xZT6U q hd'RBpT{>mJɮzȥ&K}PRKgW4J)3{).LƋCr}B_=BήT\kCgw(b6&Zv!0w{z9sj[lGnܛ]ԉ3QR@.j綞3,u#ey" Jj=9;%K,~kˇҀ SNiɧE 6 fX$,dщ͢)jXΐN*dI']T7ɮfZg+]W'vFM)[gW*;]'6nTT~pTBkK\z`|`VL47^07 zanR670+UL 훊H8ڮ M"xx7$2LCbR~Nטot(qz}5;^\h_D\˨i$z;_sZ?ײjwo=n6A,h`Bߙ{s^+7KFsֽ`TDdCn,5fHs\d=>0pժw[NfREl_x\ZuǕdhF6yǓbPKEwelCgqhVm?JE!N^n9t-qy1FbcZHĊgK-Uɱ9kt;yf>\A& ? zn^7c0H7̐g{ (\=9(B #v&AGH̐Q$#qtdR&qTA{J KX4dو0#\` K.R’!5_|xH=tiz.Ү'v&~13q&I|&.8ih͝p62753 ꔕV5|Nni瀻}jYϟװs@>Cu{b| סM^<π^m#B,9 [JdiԴzzfE)>6T'C%=DZ5c6xߋ'}/JX3*`޼3n/{6%uơ6ˡya[ mfW ^hZp>G>}Lu՗w vnSYԙ.z76}(rh7%oyiM5z};\>pۑC;gfD_,܌eٷEAiJ1$L1wt&"qd-D\ycKlUŌ+޵6rcٿ" dh;M~IQ"ޕ-G=eɶJR*K_s./y/smD2h:Q.Sǽ@(T DCg ppĀ[+ ;)&h17e)8&u4e JeH!8hA}i|˗6V FvޣF9L-УDD+B@sF P[y|_j[B0#tv[Ͱ@;_s3)~̫ ?ICN4&. ZXX㰚x;\5K7YKEcś]Uz^["wld^gs V xdV'2֕ٚ>y˥5agz֦]\jB^FDx -Ӊ}G6E*uo-bwkB^6)3]T.zv*LdZ;mQ͸K4La N56bJLѲSI]tI5U@c 8a MwMDŔδ( IMшr-ΎuZcc]v29ID 3 0lchI.ZkgH59r~̗u2t2wf6~2_sOCgɠl}1Y\nҫW_Lzzwsz1}i[+֠ kG~ ]N!@2 ,#H0Cs2SeH ;"lҜ  !QEdS3_ T6IbU!9*(FsHďW3#Q UT)?Ewc" 2jLpgQ ]J( M4!(Ӣ5ٹo_9t`|} @]&+gs Zm:[ _T|q'G|tͣ%5ȟaa'Bۅw oMf>fnde'۟ȃ;kV/漲҇y\1i:Ͼhn]kww9 o/(Z!QSXIw8cLj1U"%pDbvɓ]r}>x5;\@.DJE6=sU׃@ҡ f^ QI Ao m-ѱgTTC, *@=cRJ]R3w9]R MHg%ɮ- BE܀H^βi)`uk d,!2O*>_()\(zi1a Sqo$?A Wc̐wƆ& Bx[db~LY deƈP$πOw-nṣi|HdHKGaqUzz/E'nIc> p ͫےŖldlQfpw(hJ~Xޙ3J'JwwՂsFY*1Lv ͈OtJA3)u$J&b X$ NeR9l9` IEgv0v!EYiZ0"C8M**K2VBbՆ4)@LR#S )i5l|jfѥۚZup?7n'XԞTg$V [a:d&w2W] Jf,0_ֽ½{EEw?r]7nEɻ?}˛^ ا ڴ@G5Ӟ A PZFDnP! k}2#rRx|'Jŷ"#F6]O_x7yI&^_:(h$uX; c x[ ~7)Gc|zj %7{ėѠ*C=TS*X%NMw-RaaQ*!T#0-Y @ Oa4c)E0Z!>M6%0jI$C603@&02ԉ04yJTj c,Te6>*ɻԪadqaԸqĘMİW `c*]09W ftU#%h%Phâv5uCluZҡ}@1+\LrT!F[x͖*eN-K^- hB  CFi h%Pۛ)K?ݻ?+oFK01)9cJ֙!6"vpW=e(]!Cu 2ߨ!(Z!VއXy3$g/2գo{ʽ&="}O6&$䣽{Ғ6MfjA䣽SX҆)H {Yamԝ,iy|ҫ|HǍ;=n{}c qKL$>&ġgM,Rz.q'bcQfH,ٔ଎t6 ټ1QHѾ΀-%<]zC( *,˝ ibMV Y>:k"V5@kӟ15ܒ n2I׃;EAB 5Q(KL{:-Z +<1N&Q9T.LfVScC2h0=ZPJP&u8觲 "Yǒ,M`$t +.й(+^h0%E0NX&1T$ Ԅ+y<s rUE\9fqi1 Bi`.3 N'!rq lY& N' g)2*\YhEy#Ev6C~F ,ܦY:]c&Q`Ȩp(v|46Y9a(L xwA$8ERuכG"SD90FRl$X`()zwH̒3߾γW)ٟ)>=QR#T+9(9o&HS|/0[/%Sz?nn-ADLn$~O>,'>~~nKs'Xm=}w +$_LfrtY`A$tq6ab2B;xW \_^}jJ_LןnC*B"fT7[bXmĪ3DJ2(t3XĉMSX3'SiV@YW\Ç";DR ]p 0SfR Y;'S;Y6TPpZX$BrLtaEtO.*m[ɈfqEq-{7>6RFr%)ߋ}c_V ¿Xj{xԽI.}Sۍ>{OL~|5/!6)8dPPPx,oE uۿOFjP @.cC.68@ZK Iݘ-3w.Sq禆 6e2Ip: t!-I.N9EU)?vV/@#:~mSϾvvS-zު܁ˎ~]Պ Y=̓:}1j=GU=qK !fWq;tϻU+1Wx5B6_tϛ}c*֩ Jˏv\%{S˸&jg5Q=gZ( 1˳dzj#b]D4c;mng yvh8oyLnĚ_&Ea9zx ǯwׁ#e8BC9\&SJ^yYU0}Cjdz>R`zj8JJ"D`ۅ/Iy@gY\,SPXϓ7K.Hŝ/ xq7g?Yl1Ln@!]x`ARc%g\BF5GX` ArY阯h"UN8P)h:h)L [hsDz,X)wwJNbИ ^뤤j=xDǒ2t KrL7THyLGz=FM.o$Že $(Qe&UNYM! e CA z7 !Mɭ;P-[wcF^U 2DD~M]qJ,yz(3I=`c_՘|*:47L+P-0} p8gD(N[96q|mf@ :B$Kvn)A| |pGy׀% dzN+"0x\:+-C,ƫiq{DG1TKNr3f2 !\YP@*Dpvm]t-xE #^*R )b#DZ?#<6 2PF=S< T>a` hY`D#YHkFuy`Tx =0ǁp %5S1q9 FcKKל4]z8#XZU"c9ZN9,UkΐJ;qD}堷.= :r=e4(,y"q& ހep;+}/ь".id1<`V8=nN͖q>o4r!m`:oh"]&qnwwo xC~ÌM9mG{gﮯkG˫}uCQ,54'(O%1P.FKr8}2zJI RNjh).F(ŕ`msӜ RFt[zQIry+РNRR_ X!ie {JT3WFF7 $9e I0 {ch43 a>Mkl4 p}x$b-&װe/%P}ŒjHՋI@4@H9' ɱ˽ww[ iA%@&YE -zU|,1AmU,8n ڕhׂˤ]cxQ%&?'f*1S@ Ҵ[FE@YZX1Dž pco R2d:ø2.Ȍ(b2`P ӚpM;I!yV@Ofd_˖09f/_ՎL \>[&d rk oՠ"_TM$۾I,YmKϫj:[ ?7+fzjϮXg˸.h D?]8dju0idYԦGBbos#֐S/1=֘g"GmuXq_ZJ ' A`Go?~W9iH&݊1հF6qak w:8hi2~~`W8Ld-nq{ƭg1?ȭsk39;$B-v4l:βDZUUwMd}:-pFK붉.\NKk׶Ҽud睳ih☑8;n4sW{Ovϵ=>o+ K휃%4tB%eO k%?7ZEDGg"#ALXKJc㿾3$@-ME?|h{4B%,^f8g_煙?O.5xJ\ {Ww¢8~y|>2ϯS}vV<]ņ©alNz 9; ^ʯœ!OK Wzʵ8UrN5 <37ꮮt,PJv۠NN&Pt׍ dqԱ'Kj~-?&/V|sw"Sa!caxcBY}R^VS59!}iy{^}* *Ώ`*b Oo an"~)\WU*34;Z]їCu{YkIL80F!FU)#Bfˏt1-_~*CXǴY|9k 7G0nMX}4p&4mۆXbWYak(+4%>8fqa}so-^DzdXn2j1< M9*UuISQXiW9A'}3ĔUJµ[z!>i WH8]&WHX=V\y}, +2PLWcY*pHrt8 ,Ԍ*/tAK5|A61S#|j VDbNVbUVgq+O`ZCki^[N9EJ) e.a8; w:I'q(-.vTTJeGTA Asm D(ra$IF"L⠼Δg JX‚9e0fk[/AY(uJ5CNc PRD=lvS+/ī:Cnk>p^MMBvӉ7yVK)}ki+q4dۅ/C~# >mNW;roW<3":&'@ >X> fwߣz/?@vx4q|u*psuE~z-WH2ᱜmwoHNe|O_xŏ~{A ?~ ~ 狵|ZN&闛+忋ߙ.;5J?,Sx>XB[d眐ۃG# )!Wpw'ǒDD@3,yQ&26 U.hi c J3 gh2螙4-: c}YfHl c*,fN;MFKM sƤLJragvޠZhQ[S/\L@#êxreR L˕B^AYʣʲDB)7d/"_d$T2_dN3`U #CGG@/dҗEOq1RM>Cv{e̞_w+`g}Em%ʟL!П?s)pՆc:VKi3,`\<,7WǻkL?Z]y3Y `OO[<D7f,砩V+h8 C;˾dMΧ|z!O\뙨;o(.ܗ {5rc P K@ТϺR5?rXs ?yƓ_NWjLIBNT.@0hǸe3v8 +' `>Y{8玂D lG +30LJm sGX+cAF>xPFo yjj-$nEmeDpO<{&w1f&@n`_fKKlep&)@hθ1 w;}˷\it[qp!݌P_mx?M-h&:]$$W@5  $!їuYP9i4qNsd$/)8*{=xzdܶa~0rLaƊAimA"lpJ v09\{)m.qӟU9B> lT`wOt*)B_vM5i9ozMTX 3 1cF6)9ntjP$0 Ś*KD(EGkɟSF*C23:ø6jL/Y!S!G3TfT^@|[Rx.C/',='O S Ut$xq9@Q{5}YK!m_TMx:rOV`H{Lb#bnX aSܤ,Σ~]NnWKL"5݊[ ͈PJN"ȅ!dSC!%43hx`-C+UMyaZ3^B#zugxvcFb}9MV JZW[%l%#o{s5m)X}޼!՗뺯׻tO˗w P x}@r|@3{}RRP_}_ eCK3~xPY3D[vٔ٭N+_vX*7`3v^pGMVv`f*xq` aGBW!c3! FȁtqN!j*/^0S^zF"T@9"]%٣e=Ѿ<ޕl뿂&o^k_'Yxb+vf4DVuuuÒc忿S4n7&JRu;VC}T6h7Wh%~u"dU[|n'Gx5JMKZN9j5뒅~*Z!~|Srb-kJb\uK.)u߾KR eIIǠe!L1K?" H:B9kEdUK-g~8JnqԥVUITƁ/!xҒ+"Wo䷰jm2TO3# ''o^o/YO5!V굙u?=zJj ;(M.OTb.icpٳa-]M݉&2)h)\do?4jĘ\.J:s4CuJ&)Z@rJKKJK9o N:CAdfz_ˮd]`417a!5A=0XbMۧ?븵iz,ţ,ܽiYÖ=Tcis:O$WlaPamfkx&oB"wz9\,RM+ڀə fa-|Ued]7/Vć%[B7\Qi%Y?4$.yas|+DxYӝ[$/ΊC+$R yqS>y)*l9 ]zΩ\rmz 7 4]3-<&F\rzD9X2$~ }kBJϹ[:6sjmvy^V˒ynow#8)BFnJNҹH ?!fREW`SsQ&1Ìj|IKMѩd#70w PaJG/jA+ЄEYLmX.ciЫlbTDk  I`Ny !89ѽ1RnOj?];zH3Oyd {f:BUze'%q^&-cݙ32 6d{S.`V4cu!Bxz[>V69{jyX 25 ;3},}!D`q>k@@GnC8C !x磐a+n2%OK0bП* x@q TfYq_]gX:g=@Xrw}AezCAO.SÙ~8gpݺ"S)X BiDҸGrI1*PaF-V<QߞѨ]V'lu>W@r{RFthA!p;6k:3 .UְYi+j(Y;p8Ƙ~~'@ɨ~{8~W%Wl©&tʇΗN[Nk2?^f\ 7vwXwVWt ͇6w݁޽C1QA ; HAD/SmʇGao+5ax#+C'򟮽8O 9&XrNHG !N2/JC[?Z@1J /Ǣ Qu^l+J7;7VUvNօ;OX"N!LX&ZWo&_kT7D%@d]wv3h~?mew4ﶻִ_9YGoN@n/\vUwӿ-}yoO~7= V7M!m+k u3MA}McMgh -<&1B<,lFQS Djboh|L{0?^s=lnIp)HR,~cr#PgD~f&&MGS-pCڅG`&{bwLJف+7m}g-;Wpq#<@׍lX+-ZP"~#gwߜvƋ}k7cۭi/ڱj\@ U~tF66qYo0l+~g/ON70}p'ALtG Na(Yn;lwG}7Dkgx9r$$x4[O=>_"vGKvA`,N Г.2ѻ.8)P&hƋ W/M-E~6SzR~}w| =鷾" dܯZ㞧'iQgzg_0* ﺝ񯸓_w<׫%l~pEwG'{ziZ5{I~O_Q; AQ\70ձэ= -@p,aoX8&V3}+f{Ja:Gd9?Xщ}s ٛ,cukԁXio"%RL(4?; t8_MR@>Y*/x86nb8h +<gwecuӐ{c5}܇{<<4Vjɫ?%yqz.{solmU} 889U"a!ETYO## 4!"8( H9gꏇw{oQ仾ZFQK!IIzQ]fǏd#"XlfL uC"2$T2p#LBm"10DVSkTK?A.&Ryz1̥%[6NoSV|\kP .ͳ6" &6)BDdX(H@'0BFK ZAxzs{ :PP ;G٦(d蠕 ڒ A䒵Uj `WhH)%.RRhN9 'Bj)TX+A*]Z^0*%.sMRzVΘFsPԸgb?N[;eEBckB1"$tWLT>7>;tJB܇ZA:ORXf T󰐩29_2|"i+FQI h?CXP<$#SCP UöV $L+T׹_ױ~PxyW8*&1+(.Wӣ"HL#&nÃ}y-πYGާT> ǜ8I<*oF7_t`u:'m '8~ ]7 h!VU9Ǚ qN1w%4uJc>IR]?v~ݹlw'۝'ۥg?Iu$c4D`YhZT,'PaO`+IA̳VThSgdfP)*%߲HIɗY f60!;$ 7X!6 N [DNhNNPp&EQ6:V/Ѩ( ADA# ?E5ƨ Z 4za:y[U/+!J|;"MԋYO>[8[>"LcJ!xqS(āD1L D2En9OxhLXΗ'x6phy`FQ[tQW&x#:+94xmύu=8<:,YU԰r(˰%;)nܨ36 ,@G6߬Jǟg^5VYZdoics@dkO矏~GM~4 /OLo4픰Fgm:X:/}p]XK `.z5PhO7iӝ+hhI,EG9Mݚ&9ݩ.zfm ̀ai+Bsk'_ 9*8JeRFՖ+i}AB,Nf˺=ZZ% ud%vGZ, s74cK%84N!ਠULQVQ`8:HCAٹt9e@tP/!`<3i%b쪜JG&+e ,uYym0}=| [j?L]YuU6` ~uϧґwT%F`C7,_r5ꔁd͛Pߺsn| Z_~` g4B!40̺- [(>C'&֭?OU[Ӻ! SRLUbuǜ vnC+5dw]ք˚>OwWdTٷőZՁC Tq"/`cf`Z;qovX:a @JTg5ܕk|cћ|F-ϳ;3[ݓ O CAu6H7Rt jT5)9bno`թ@%Q;i)=6/x5VJ$eLHu3z0 ݌a8gXn~8$ n|o|pωůsX&JjG Y262SJȍ}f[Kp >>`9D H9ȭM#- +I2hRlPL_HpQYKX Hqn9Yrl'ʅ<rD!|nZ^Ru;\_s aW-(eWWn#k'CL]_]Y9r?{XbU s2<{֬eoV_o=R1ثVX* A0~`ѣ~ O;Gґx`)8h٠x8,s+&…Kev'l}R"@" eF)N+ҽ#3g<rkb(yD #c`Қ"H [ 5a&(NlVÉ}yavZ'Mzb' !hK n(MH q J ѩ\3NRi1C2E!QHJ3i=#!ak-0)4Јi+[&Y,k^!/`(OCBHO*^[?^:"( \8-lCz~ąn}&Ma]eVh .$vcN]NXub><-o)OI.ڛyR/&ޗ>FyϷ*WZ~?}޹ʞ@CbDH(OL=`*&DDtwPʥNrpȫ؊8JBf!z(ԴT}Xd|(<6W?hMz_ {t0;Lq f殐WJƝtqnb&5"䓟7v~yc獫~^50ԁ*Ns1oVvCwg[w5A] عh=P FR^ &E\hNr R@02jabҒ $R*]La+{-gU-٢3K?,S2wAc/8lB+uH_n'w@PE7F|);~}rҧt/UN9M.lܺex"YA-wlelى>€v~m@ ŵ^ ״1h[ȯYy6lRP>}z_(Ocr}/K1P QK VjUX:k%m<$ESV\AJ0tY>~䒸{AIs‡8˓Q4uzc9Um淪z^fjLp SvIjׁ¯;δ ځf8hU8*d!aq^C' +u"Ŷx{ai+‰kĕ#"*tHG5 @ AE5 6C@qVh.- vU,O.n烹?Աi 󺻰+q־yE\8cbOunuzP%jGWsZ{] ]o(/1ܰ\jiF <s%aLfgn'9)L Um!RK)!v䔻ζ jI_uz&8ŽRZ¤Y1!f:gCۗt_Q"p/wuѬXn#{3_ G~*3g=fѣݱQB%Uh_=4q/fYtuN9;~ \oہiMTn|r_b9W9zʡ\fUQei|C4hLI>xucU[(>C'&֭=iam~.ib|}əi uǜ ڹ;Y 5&D W;]j$5%zX?K~5A_hG߾"؏Aar;:%J (Ψ`)aV290|;Seh s pf걳ߩ\r6-2<Ќm)-)&IȬ#X흗f|gFoZM\N>lխxL0^tpTQ1T:/H+jP1p[5ajb PM}瀲8lr\AԿm"5ȨFΫ, )D}Oߪ^6(:<= I׃x=Ԃi6AYc?7Ơ?k1++ ؕܖnEuS?z2FݗSjtL>5z}YZI)"䛧.( Юg.$#=L,&a1گ*}/cS($35 b8Mi8̹,5YRSaW2})2[p$SqY^jy.tBq:ilꩌB)8(jM@.6eMHHZ ) O Ϭ@lݼ/.PA I3߫Yi"C-E}v"NYin!~܁,AK _QQO=%e=%D>^HyKoNC5̷w@.! e)<:o.KPJ(KEt=s oF|q[nUwj^{V_]?t^v^cC:@;R}1|67Kف*;;~ADUﷱ2i&g%F_J߮X|q<7`TؤbH8Ӊ6,W&jY؟ } }\B_A/&ѮJkkݩhKC#hS%f9wq.rTB,1K~jh$.%\9oEO$|*i0?A7"iJ)KI0NvgҬe2D/p:% 1Kg~/vxgKyP?x?=\v޺nwG?c3^Cjo<;vqće0?3B]ε]纫Jw~[ ^ 55)0g#R\P@`E7}㦌2aC`=e6t)Wfy6ŭ5othJMJ6nʘ@o(r +3kaubYh~oW8w?E׼JgOgN~k$x[uK⇱;wug6O;?3|'o|E4T蓑VT\f+5F-_7|ZhgmC0ԒDg'Jc=o>E{}]:x_~^yrep=_ȶ/LGG2z pŒѳ3ײ˿9B项0w#ɰ<_,VE@Jl_3mirBڜw(mnKUAY:; l8 PkfYN&r 'CD.58;8I ]ǽB32볭]KtRjm .X@DM;烦V`lZa@{h9jIVvAmWԐvGfj+XdEs5#%6-olꭤcW>FC0gk*W?O6Brۦ Tƙ,܈1-5Y8)SfXESFGq46a(3Ìr2'wp VP*y$a$_9r?EIje*+|2h˛g<+dm:*}%8 n P~ZcFqx x`f=B= GؒN Rm`&dnB3ԍ pxi̼y ԙ/ŕۜD Sc}0/%@f\uBM ":<кC2<.CMTJ'*\AuL1PmLPNr)J탪hVj(Y#1t>5" ǚaAU7[c`7F5Mqpɑ\/ȜIǠ yJJeA"c(ig3r !WP PE 92&Ld  e'H}%P-SN 㱳?,FbcT])W* JH!j=l9 |ii!!.W_5٪F>.G%yC_rjAWnVcDuDfw%raDE>' ?B@MmD,{`[p]RX׬]W|*W,(T +[]=oF93 fsrݣzl>ENU=6@Y~r\*ܫ\sN\Y-JplBa(8@9>(@eL4Nꆍ^;B{2p2dJ<"Kк')s,V9:&15I ǫPXvm=f&W|gJ[烺 eK*̗~߾qZee)2-F !AQ+iT_U*CtWńi1voQ ⤱KIVdLP5 {nQ!,%i:U͸r*N40b{Ÿ@d:%L`2YH5"FFDT,],1%^bn1K7:r BDjupV h(Iɱ&Gl thH2y3gĭӂ)A15SWӋGSZA dFaEyQM1/2`7C'K!/,YͿu EٙY^٢|:!D_c4 -J|G%SS-kASv3MK|?*8gCNzZ-} ;Z>@26qhg8@} _뾪iMa4w@91T *eȴehKGXJs0pr9A3[7}G> ,='}uОSw~H3LZTOkȒ>1|~l4GP|1٥,?ݪmD}_?}FC8f" JTݒEsgwE 7=U._ {ВqC r9cV➭zB5xKwe IJlJN4 \}_SJ̢yaJ=Z,eo\\ -|,X̢zZ*c㚏E\+)ؘcm)UbļЪ=5GDk9Ow׋u"n a }k:^׈:1pl-M< E}FQqON&yE DINr zJ낳uV gLJ _mSgb'8R]''.$99#ebBeJ3&M&Z0L[IW2s4N_ Jo.vwQkk)BLx vh6z[PT>RXOaKUhj$Y+_*b|#Kxg>˫nH1V&Q<6%ؿ伇Г-4V@cw3*)Bh,BQ-1>Ðd]|DOPQUZ#Lt܅DHc&%DŽp&5J/U8T[^7fi@ ڠ rcN «s&øOҕ_NƻG̅4x%4Fau$E Ra-)㚍LQv깋ʣ=T^2ZI؟g(VR2]2`MTSl)/x DJ.EԠ ʧm*&˨}PYF6_0W%浘meU @M?I8]GIy V!z˯A@ F6"(LS3F%!VUP܄JB%q#Q.d0ds6(#m^|Ys8FL3ZE%V|nG.y!V8?|i.qZ4]W]U̫-(a!ZmoE髫gW`ckT1 ;1n]ܧ5ǟeO(9P֥ݖ sP1b 0V0`ҷoI O!ttvY1&'\>60xk<>ne_ìOwo~[Q]/aɵXv5/'98b3WTj%:p^j $z!x<;A2VeԢjdג=U( r5)q/YeB ι6Kq׌3"wrzD> o =H!>޾hS04tEKIdZ.8]s0Hְ`&D*%A(4fPj8n{ŷU¦ɷogD;kJKKhjaԦF6t:o"R"Iyl'㍶[O{ DF9L:<$J%-@kyDhV1/fz#_i%ɍ[NȆ$62^^b$esnIMi,04UX,R@ﲌFT('eZʅ2FƢ%@u+q0/U:7MX2ڷa( h ¡8)]0[LSdmx5PhGM[+ "b7Ž`:goY4C:r.؈v;Fl*õ%2p̶>\w~Yo-3.ڰGqF]!gJ̟ɨ+тN.-Ԗw')I !\?*\ 9 bN9}2r|N>@G9r^?y?iL:&uM,ZT^Aһ4Fl)coX> -crGZ1eIΫ>?G3ۋNvcV{b^FCU RF&`fsr sX`C6`Ȅ=Qa׀@CsOeu.$Iy#zU4gYdrpzJd+r;{ hukFzD?~p7}}}3h ̨p1E!c&Q<CctŴ38|Ť-5x8{\$3fkC椗h kZ˜p~1IRy%7̙ \Zʆ!8ѽ˫)7ل2Edz&]8UXO.I_r-$|jڍ%undN`#eڭtڭr-)<~GݺvKJ11hMBڭ<=:SՐW.E2%}kM!> ۭ,)v>D!ٟV竮N;{qa'z+%͑[ 7~e3h)u+ 3ṏ)ތ8OZ ‡͏6WzK$8]R\ւt{;ǩq7>ld~"uegS&TOp2sG4%~e-$n'q8It7Z(ϝn߼ˆPA#Hlqg UTz&'.ҁx;-C&'sz&''ifǷo~WoN͛od`a֞|^/& vws*1lOELm&xq~&~ͷwJ!M06 +&-cY9H,ǃ^3wenHs̠C:2! \<|8@T f|hrZ yp*8|v3P9j27O6<;G);+a+gzg]:c O.ph9.nn 0mxFC,$/+[vjDxel͝ALp!"(m/JEh貉&0ɕihBAhZ"0D)\jƂhSfC$+?#4%vH5A 6-`!8bh lqh0c3`8pJ5 %Pc)>" uYr-).ԌO'х|rsTw&jIs;w=7f߿{-4HEw$\^. c ̧L(_],6oWtߜ뛯'ܿ\ ۋ|$pmvFi8 דa $`K]KqR(J@TMح+~ٙ4 }kJv5I*xٕuF'>$ nlRCBKRtǽ=Xw%s.N\aGק^us\,.>|"nfߴuUpZbW؂Zʑ•tJV፲XI25rJSw1 Ř LItBKמ#*:9AݫLfєG8'M!9ѕipѭ6ŃPTI-؉CAȰij#=DŽC Fs'E8 7k{C$f&ɆaAuKT5`EiٻFn#WXrˑV龬7Tr-Ov0Ddޕ58$$ SwM@F #Y&ɳ,VUD1|u .G"vE=6>`FH2ȈC1h3ۣ c]u:z $+:YpxTr$]c*oȔdCit K%"yr뾢1*OF $TV b%UxG߯-~?BCAPy@%?ϟჿ৫5F#|d/V#ڿU3:\}f=0_̸ J.X7}s<wq!}:09&%Z\{{4mԔX<˙BUc!IskDVLG XbqQj{2UlMau/uͯK7_hsw^|k_z@ӎia:îLۆEM$H"`%3иdjD;*ط&*BDs%nك:?J<֛tHH=~ wt2v2熌fxḣ=+ j8k2Df+#"_g_& ώ1ΜMXKOfm*+2:\!ٗq1Π sS;™H"1>mzά+V` F̓G/fU-w.6^..Q`M~\k:Tb/WRiпH>F$&X:uhsېڐbKs!Zq9Bq%g#UEOшhMHRj3Dp6rS(rЬL'ALk}?A3qNrJae. `1j=*C$1= 9`4-1.!-g633d,ЬrIs$vVC<4%>ZX1d)ʰ}*T밿Hp{3 `(àMJW9l( 7F"e .`5L\*̹L _apXjCYlno'`rbyP|\"ڋutXzv gI+[jtftHlPSCNlwkl`r*d߯&;Ho;BwwG k֥ :S`d_Ax5êQh#mnde}LZL4)"JѶ9b{ɲ֝3}|E%-GXifmΙYKEKS`zȭCn5ղ{EBS>d@Fu|S.Bh\:k)ㅣ;6QF#mW hŢ{g/|iٟ=pdWgwVvmm:8X~nVmN]sKI,ujkj3(tbœ(BoMAj]n͑2CѠ~lT:?=`&vȰw=<~vk|80|jUEʪ> 3:\Qwu}埛OP { y9(%)+STR2/%GSɰPTKz_Nr&3i8}+Բ=5TOf~wsuLrk?Ӡ+Wg`e⎼++74C92or2\bN*Rq 6n;} _DOt~Fo+p+ѫvoE[i"ewH[w][_Gwx]Vog뢷M^e͘"A߉"^9AߌH"Veȷg鱰 %"o~7Rjaҷyxe[NjM/S`\EmB-[˽S J$/x9ҕxT/v2W/Ȃ0y!fV YxK$o·"B.5{9н0KZ2)=h tDBy EHt=DVB*=IeL!EfynXsi U:PpE-ul̜KBmp V D4CN"*Q_@r)Z̎'Y%$ZmxO*(vJFj8}Ǘ^"Ig4M3ncCΫAϥ(5g>b+*gg V3 ^f5j'',Y:FQ\Hֆ0dԵ {J pѺs٭'#Upb*=^S$JwKL$VE]E X$~|8Hm$*N' *&L&i5pS% / n3*~&,l7Q4+v.@9DѺ(42"+b:,\[d &T p(H,9 XF_eHZ#Q:jDutfXVbWUajBp r"ՂZ0e Ձ/0`2w)y-6ڛ<\VnvQ@v 6Ÿ1fֿ~L|1Vo:fQ{V,[EUq_0@>zdE5iA햚;~wZHBpV.J&7.Q2Qlk7q[,!;FvxYg->hvK!!o\D)`1(OQN#`| ;ɺ^/EfmfCək3iK?'E';T}J *S\G20­%+Iܗvl-MM2J_qܓtHE9xgδI,9夻!\ i0)9tȸC--Vrdú_ܫVpUBzuO {&E^K9"ː+3c C@K%fa3[Q}w&;>3A, )Y[x^Lu[=VL*-Ÿ[ [J?%灖%/f K>/[%$"Nj B/IV"=y?q2Czٹb4m9oH+ZO岛PHNyŃͷpsup"#!I\r1ZBd ,ćՈDHKabJlc% %.^8VHcK_7ZB-)h ib*f# lmV,+Hp+(rİs,O0FX)!iQe!U^V:Ea-r W)*`mS1X1D/-/Q`e>0yg21\Ua_^ߐdk÷/ͮ*feO燐1yx6!F'_ t6_~¢eG}mYC ,}?o|)f¾^|u @;=ٯCsts,Nj}<5tAgU"%+z/Ie?5"Smu˨"6.߶܂sf\#{I5$W 9dN!#3IDԠl9Uވ8(]_{SB %ќ_K#s@/FNK Ȥp@ 4 1b*T:p1/?w('=OF?/ xr; 87i?J+Y6r[$5~$c~cM~t`5# LDm Wەݬ Dvlj3ռ/B׈ǫ OS! /M?i7Pa[&*4IҴ4 DXb.O'Nr8DQ@1D͢:{KX${,mCG7hpq .ysX?Y}PXA1' +$7(d”ߠ܋FV2IyZC% /"Ь:G!7 1^:?μOo)#Z@O/8C j}_."'=.`tMRvkp;ΘX,jGw~{s37OnW]>e^sahF01&ιCЯڝĒ0AْYjk/v-@ q3 j%JÕ kԁ'v'Bi㒆KQbJ/8BB't@JL ҇ ܺ0QwףO$!ab⓽BYy@!yL#б[4 r2+c[jNAav /VŜo!~Qȯ6߻>,A,#me9Cx9"x o_#x>9Cy>.m69MhÍGO[mN7U c2N`]lwFcäpA5 矩Nts[kȜAg"_+ݞ*zU6YV#sF2{yuҭ^=ch8!Ңベ[m2%n}= 7Z0nrPu{= *"I= ˗i9 hLlӣ]Dq溋nj{rL+?{y+Mcj28keRP[]RS摤gfb8Cm|PEf5:(όMy;3 tƜБ@$?ZǵǯZ_'V ?!0^qH\b$<}-&6Yq2 >EFL E2QErQSCxcP6|%( #نLڽh~5q}@58JO.@! ɐ45N:[EÙ_F [C"k6 :DԪ3R# m) n`}k^ Wմ1%&ZR9JE-*&%GJ"aK*z ^Гr`K}dWWJ9 $bi[DMM8p)M]f,mMAh 5Rcdسx&kT6f?G?I-]u>[Pb:l6e^k,w͚+@?ZUvaܿ])ݬnfݽm|{L|9!VEdX._Y?eCkRD3̧9(YnTpwAmBDAjm@Cw3̧h9e z7*? &KI3a~Mf`a+@3-bYx{-(4CWj,ҩTJ2H !Fh+ip?Ayデh ]"1:Kn|wnWUQY?QiieaF2!w|"q[, ElbTtnf[\aܲ6J02I@ \6̂Ju.5b 0JBdY&ưO3JOˊR)(xPvPwD  oˎnVx~M|" D`d>x!!8-g|鶍0K,FM^D̫zέ%C=CAHwD4nOwr1k3Z"q{+{8oΚLۇt|ft5?>$v{#A&c'p4}ɣT@c%?})N&,Rߕi0 J:/!$XdG 'BdGB?i!+Jx٠D,pR ?+;>hnN//V7k;}3ǫO{7e%T勌{wE?ݖ7O‰7l=`KYwtf֏MY H۸ɷc98ph$g ٰo԰RMv*xXB–!)ªdZVXLIW9K&JUJPPfM G\F#5J ^LU<QMKHzG?N!҈'j=_{qR6],4<3 g.6K%4ȓRBG'(]4.RY.0TB EF_?~ZFHl7lJ2J`wg(h_yj2GLZ@ H Y'% WuI R+jLH-(cLc=Q$ǑHgp1" )(c,OjI4ǚ$G˹%{~QNSVJIjD .|™e!AQ,r8'Ap?"H1&a]ENx&*%[ NC!77_Խwܲh{= ef"3j.Ph=rgiy? Uopc!I_ϩm8QN)M^\a' p3;gb\9B3yvj'vfÝu 66/,gH2$/*3G 36aK&Ў,7,&>.vP+ޔ7 YzOK Ce}@5!O{a`:LR1F*M.(5FהºHcL]RW *2ØƝY7r/$G **uyk;_Ra0Ę"uh6)ט΃mw%_7@?@G09%ִ^SgPG&q$LlVDaXH $!nQz&RFbMe!,6,8':)RP&[5.%APJDY-y = a6Eqq}QgS@3&  ƼwLJ9_o }N]A_E˧caUx8g᧿~kS%8ȇO_Xx~; "`j?"gӏ.`/||L ]31Sx>,5QHo|/c_ }g7ֺ0DZ<CRn( |ޤW \#\ݽLKp%TJ]rr/#ͳR4ts,C_X~Ы֩6MA#As{뼧fC]ͫm 멗ZI,l@UN,Y< B{0}gv@3"2ͥpoU…Ca/WC,Ve*SNA{v܎ v"3bWl%† OHX)8OV@,ZČТ9A҂SIbQ)O\WζbEqq %oJCv݌#$!WW]lnLoeSf@DQHi gc~-ߝV=ls2ݓ7cꌠŽ96pA2Oj$Ok<D9F}|}>%o7#@!H =xKUO2ޞIH MkaLayS </j{xFO}1QI-7D+q:psMZP6GlVXVX/A-)6%wsɕXlVZuլY|f7@2F!&lT?tzp`f..{_U:1|{'<oq<“58/',%%-BXҷ2.{YӼ-\{sv= =Bx|+wΥ`D|_yK{'uEt]l:|aS&cZW;UZCpU @ԡoP}x_ބW=ΆB>^Y#[cXi.HVgí%C`̮uDVFzD*uTskzoS EnѯQ;Y֒_@e:IIOw3ĀLʅ 4񂤜[b%Ӗ8_߮Y᤯Kj)/33` s=CX wY> sSib*d=D9bht2-,Dh=s S|0+ldj=)SJ {*zMpaR]aaJZ cns33T*n 1o j4薁EX[,`a|=Y]t_X')XEͩ_.V__PP/QP&t^"OyU(4HaqJ[S/gE㾞ӿl,vAv;_h'V#q_- G˓>U84龺G6ЈˁdIN:ʪ (hD< vI ʼn{dRgL´ (KQ)Q$DQ{EJ:r<ԢZӾTRHIU#Sutpe$LjϥTZI5Ɩ mj|X*bLippv0#փYkAǹV cް% W-*џ߿`ޛ.{..3ٰgP{-)Ÿ؇ǛR%JI_{_Ԣ=TYjAs1%yٝ]p@cR(?|ƠeEYe)l)(bLp>ǟ5_WdxC 7BAo&䂽j~Ð*= )9t8! e-c.Y1$ZGPX;dlPAx\h8qxo9\ .i1qMx@o`yw4i_ho{R 76Г=;9o_RY6*f<;&mqQ1d04 輽\/?φglRI]yk{ -OFGN6)Ns]Z[ LRᅱ1D03OPo<18u5hE (. ou䈟}=0Vb`M;xÍiƖc50c&0\ny'IθZ(b6A b;XX3IHC9W)OdJ\;"_l#J0!V1̃AeDh4$% jV't^FD s>\\`7 #)0a mH aBH0&aԔÒ樒02dt00h$$$8c,Qh[QRJ5 VҨ4&l{6OhAte,=Qa$vYWtgG5mv"{}P$a߷JpJxz/[6z8tAFCpr?.G[[ٺ[w79ؼ,(7oh]霿隻SEgy+/_s/Cw̧P yh]j! +{TUp N?܇VQb{o ));k:BWiϧ7ii |8K[.VvkϮ ={c*Df"(WWܼ]X%gv9A/ ,yD@?m`k<BՖGG_F)H̪b- [b2ȸpԱe0O pθ~SCՙ=xjՁ#+DM} 9#+D v?&u`KDnD&5j9}(ɷS])u3 kn\WL)M& jSAN̠[&U_c&r"?=_"qw/N.DϖwP8؜I$Nu튇Zٞdxm_Zf3_m R ~ɯ)}ҜNr&*٨5>ɯNxǍ2oP#6-vu]}~_ivi+ ?}U櫑gO=O_`;m |goY.\+c먀XM FK#1%ꡨ[kd% %5΁Cq@jP.(a<,5̇tf=9v|!O& (zI1lr}ɛjsWe/2o&!r AhD@YKBb}/4-+TZ]KUpR ox_+0_Y_Vf҆GtkC]SVVNZMp,,{h#hɪAa%=9cFw5y%=LSI2iH֓U!*wj`ȗ"_jR ;ƪܖ7mytҴ^ۖDϑ -N/U󦹽K")OJlg^`ǭVPs?pq=fOg!@sq@wjWbP(G]tE,Nn$VN!%vt\@sI^U~u$yHFQCҖ5v?EEnr ,D:<#8+oa}vj->k[!U|s?}uMYb˻eO-;jM^41>#Ǚ#.&Of|dߵ\xx{mVJ.ֻ<ʡ2nIخF7G7?΢S?: _^(<Ĥ?5/@B*j?,3igZ5֪ [ Px`V"XTm%d`PQ(EH06mkWYMSJs](-a|L]5|lfh{|> `7 0QQL3a'벹 ,\'4؟ŝO/^ˆWEv\1J*jVPRiRӭVQ9kq2[ED%ĩC`F#~9+q߄O/rjI7QqċA^_.r_o8?~x 6z`o/wxvn@@u3?!b/@XIm~\S~ۮօyN*2+2\}D˜7n&ȊRW˷o w7v[[ۏGWlpi%CLy 3Ђ,;s2rQ0G:zsv,~M,>tGoJ1UFAd;9#C&-j3.>r[8\+8+[}'~~?,oqo_-kۚӓsq3c֔s*!^ݓ2UDFT8&;^3׶&LQe|pOn~W )$Aa(1JϼqoW;z5tmt>y= ! 0 8hoxyl8DɥF514^Ȝ7UhFAJl ;מ]_<5w Zj+6D#&ci"$l}Rl$~)eO(w 0jeC*\2H$2i(MaN_!zd/1PBY\ `  r0u*:R2-eYF+GHV J60ⵧYX s,`Z_.ZCm~y9c!Q2X̶xݞ힅<~7V.F.dSugc=݆ҠDtQhwdm0m y":H8x[vNأvJFvۣFc6(סѶQuϣJ"[i^{F>}gܸ뻽X`( *7fwhE6`\EP~Tj| 7{('25#ImJ)!2߃]bZFg+ b8;ˁ %xͮ3DKlL,+㚭ŲEO:~ A:%-_I;j4譆K( hESπSXd(;i6(l%Fc@ғq(bUØrj(Qskll4º.$/70&Kdw;gd< Ĵ3ͭE j@K֖lyp!p$1m$#h SSG dYDMYU6 dDb= (cd`(9Ꞿci4tZOMO0 3'f>qF93f֠PK=R#nǫD:5Gs]? 5`%Skfgҿ}L?jyZ>йImqiȥiO?\Y{BaK:N;X3n:”!Hl`aÖ AL1zLG1K1YRjϥODž}@;R.  }3l06{`մ&eHvNk{T5 * bb3덝M1Bܭꔂ{cC﯍""mYy ]˖k"+lG_Z)QuyUa1 wěW}Ճ\ʱ`ޯE핀ΙDOr/EЇ褫B\1fطJL,Uks&n5,n 7͓,'Y4OhtE;eB&%O*'")$3 #XH@C IKЉ[|(5!6;ǮX󈌶6wҰloJm0ەZ zJmt1BTRN5ؐZϷej05DCyqͺM#1'j*kX7zXQYD}2fZ":DJ>P= "ɑFjP2ks9&ቆZqt /COO`2Ժ(d ?5jZOwG4 &V|3!Pw@Dp U }3 zy[E3]SEC}4Bг`tzp^;VF%9xTN`\kj)\|$%{sҠd5j$Ofv$ $!!/\DȔx'cAh4(khY_ZN>m y":H-G ,!h4(kh}z\m y"[d3.lq]3*)Rm]jʪº:PIvJ1(Q|.90W0>rkdujE?ݰn͏!&L?ߊý~1gЙ3kS3`dr*3i`'sMTM) r1\&nwZZt"4'ϕ( y^{HZ+é-XG8]RVBgETl05M0,|m6j vN-A4( (";<5{A)1Ji˔R[QQ5sɡ5%kQjR*՞Eb'R Z(|c*QBH0ODI$6 yf)ΆIu <av%3ǀp=D;0 Y 2/WQJZo+sFwBc4f-rB+o68}N]~CMj6]\K@(9#4ac_c&}^lJbN&sen:o\+loak+vf ?IO\(scp 9\sፑBP~ )T䤐YVb;z?+D+4DuHv+$*ʕkHW Țͫwc3s:Z 7Qb1/k_qz̺X|~]T- ʂE ,l!.8Gb٫q xK9oʥCX9697+{&7m->z2;9 8F^x#ɶvzd/7{*#Shp!#L#J;=|3Ls3HxGϕ$;CrHćO# Ixnl` g121) ZlY8p%iBE(B/VI;{8ȏoU{1AZd[r?/Zklվ->%00j=/S;ˮ&^6J<9ǒk Z3 4{`֊!7SX2켐VU'U}TyRUU]#=zp#%,R.%U(1 *H`ƅlhF ФꘒHPj1T]3ШȃI)Piu4J D9;ԻE EjVLۦKm-Nʹ5阓ȡc J>EoJiY 5',nF$cV;FWMQhm^rly*_rK;[^Z=7 յ*w?[[ h 6_DDhQu~@㯯O`eɦ7崇J_ΥZr3YRo7WWppl #,`ۃzP D(|OFܻqCRLTL")zT=QsڝjEO5Pc[,a]`H쟼5.[4t@4 y&RTEj`v wy0t5uARl"VG W8JPQpVYytn,hۍVOR/Lcaq1ca1Lg}7[4H0; #C9)o:SAR ++HdL3ٖ灱 Z32芼[';/|mq.(GOKQ5}R63dHvrIGuq$kh #9-&&ɶzgb L4KbHUVHݽq;mtQN6ҟuUcK}JgjtF(fZ",]'"n0i]j2_8UҟflXNS G*Sx^=.I;묌:[HR T`VhO4V(!1Z`5\XtuH O3ݽ1,)97u$"/o].jHQ@?{ hGE'9D,^:(yqdݜ,hArbPh%OpqZEDpVgc '+ |TtJB> `.%PEP{!Xe1e=2Z *PD]b<9O6\1!3/;Jo*8#kn ʿNIrS0Owre5vQ?VKkg@ŁK3#Ggb<`ug hVbN{ͣ)fp1`UNwnn-Q9Wkx/<#DϬd h\(La•+vsW Obz$铰@5I(NߘC͔dbÌ4ɛPC2+:#s3Kc]0Hα 3 c3xA講wW~(|ZܒRR |v*aqw\FK:HXPK892aBqݸw="C𵤮qTN:v vGhRN NbwG ]'vj* SH5׹lSsYXLXW:INn(̀gN?IRf>ͥ_NSUXiWݿYx')[;VYW~fl~4щS*h!7J07Z;8R>eqn!=SEWXH$!߹6)Ώx4L8vkJi:pcݛvk$S!!߹6)*Ih[SNwntfή[*ڭ EtC2Zc: PG[Q%.QVD @7fC8_Mް3/Pgv̓#O9O,8̭HD{"NMѥ TtNIe$X=|1Х5  )- (^FG䌉E(rB%m5Qf*)2u`2)`-T1UF`5pF kKOxϩ fqG(1ރ1 [EDƽXpA@x'M=iRy F9a aOsHQ*҆wR0T iPD+yJ#pD JEEroP@Q S)x  XmX4p9I"\9oXZy5Iک-N&^Y=.||7{WFf\(YQjEJvԌf֔&RvnM*zxX3v^;q[ʓJ)OJy\>÷m’\%ieg]%7~kҕ3]5ﮦuq6_8SXPMR?>81!w ?i`̼r S4T|ƋQ4\^!0fqy~`HlaKزٕS* =2eDTCۓ\MX"AI}JjfIs1la?[)DBy5Q^<-״M@a)w#6TdžP5.[/-|YЩsZKPcrX16>RIll5rYVŴZ*~bS3&M\D_bG"&6>'Dχhn.E^>~(bfn/UHYh{A4 ѩ.^O\&v&͏։&Tv*4e(xLX]&jGɌo-&0o)7M"\H3MX#LpYhM N>T[xmII!T@RV6 *csB S~G1Y,n){>dG|ɧ7,mI 4ь=dC&j"?-?~diWia39)kooVjeK B5_6_03b܆OQ;bt6$iL8˰Nyߌ:1N9v'aJkhidx qF/v쥷pP=lS~OI `u) Ig)CS<:[!pMɣ'Zs]Rt!]ɖcϞ8vǃ+,ҷ4 |jӆ?9y%Q5?2B0sts&;+積&3jr꧘$z:lmuhh-  ցґɖkhu`49eEwnD'֍lJ &Fm;2RBU/9"I#[Mzt.@y'Vkw')RhL/>yN'|'o#VT[u1YǺ }z%)"g>YV & Vw^j7Ja>!b }& 2M=Z s܃ؓi"/Yz+a)X`DP @АR wT}G=[s\+66@'8=<#NW>}7b; +6 5 >rzZ|{tuz8qc*!7Q}l8y}UEd ! e%aNY͉tU7Xξ<["uJ#PA]-䭩A(YdSɗP>N%,`V `!_[\ru6R" D&09]ivlpܹQd\GsQԆ,[u!2TfqߐF^iydU/cWڱ;3Pwp;cJwZV@HL4Gm3tgLӖ3tV3Mdʄ{;ږ?0/ڸė?qE4%v?rAslث6ZƱcg}]xVVoVRFVJV̵8}/A~jv/+-V ogñik}?v~w!7Mc;/nv=iAEWvl:Mnpmu]V gDh˾fpbddc1%qM-@f[f˜06v~}j;}vOa$4 y^Wt}5wȯaȉ).v6<LEfHjwYq!M8XT*\zl~%{fqFH'A89q,H iU{֙[8|t9Av<ȖPh8e >Ϸpߒƭso܎4Yܑj::LGoϺ9H7Ġּ> >e t X\fI~2c}VT J>F!eRP\`ac 0 %N9FaefwIc?䃂W5*A{(ꔢ''jm~f%BOʔefQiӎU[b<(o1Gn^:}=\xsp!v\=}ޥ5޴wzҎEz.MҨ8œ.)]<& igÎx2Mޯڱc6wsʞvʝ͌"6ab?_M-ɿys'@sxP"tl6wCwNtwWcWe (~Շ?7yEZ'qԆkG0ȥ_իˋӯ ^ͷodnX\ļ%.C* Ԋ,ן(h%Y˿?ZGfךj^No.û4חq^CωiI|?mW??.g B#}ʿ̡@y%15 y͋.`8R]'n4kMѹ:9Dbk+ͪu_8C -,j% `C7<MaB3GufPu1p4EWț3Ef^WﲤBkEɍsK|EdILt*}7,<(ab6 Ee1T;X})oDoXy4yaSqkt<8PK̈ڥH{zް,|.8qpX$ccC&)) {yc!ȜB9ש~}j<׿?rJB*rg+ڦ9]uDJN3ϡK~˲DT'S^+8lcpD&gˎ7I5P=EٳG~'rUE.7U}qMUtýU;2Q,B֝U:͙D{<ծ;vv&OsKN'0z-k6ap,*'0whrt,; X_T:Z2Y2p~?c=]hl*ot_NM}zRcg'w-ߧ fN]껂G%?Xy´2OI3tx'&~"zLL*H{pfDz Z>didďz9ZОҧ;z ButNZ}D rt{GQۻ6 zMw:weuS.?ߟrKDF=nǓP|vԂG6{+jygm5cFM}Q^ gOԱWnS]o.r5Z~S_L>pku!&o9\DjF8ˉboNҽ>L`!x?C)ɑX:T A0OK3_ck=E' T`יyCw5o,9tmzܮL.TAc^Xd# SLZsQ;16ֳO.f\6T:ʿLI5GF୥TA3ѩR4%̜!WT֎0/?wPI7y6eD|hzc;;z}NۥC!Yѐ>ިVþU Fz}YhD?a{يޗYm\lMCm{@mkYdj{=$j:\NJdj7; .)$ bctoh41k]m7r+} p@ϑ*)@ .@ i˖Zg8Oqf%4nwɪ"F2b?%uJ.`f+I|`G7HTD)"{rNM9IO¹=pKX.kprlQs,n,LA,zwQ!$HZ#=S6׼RdzBIPiw@҃KS͑z" =Gg Ԗ6ނt $5,>0x1]n[K\PRo @zBwajfgi3y% :۰Qe1rbO"iDilD21=ռȔԠpVKߺ5fK<בM=9vfl7pv)\.grKVݜnju=Fõl7;1:4xI)ԹPNFsnڥnZtt=|v'+Hg9bh}~uY &TI}rS`^cC96 nr% f K^Iף6n'N\z4v9wDN.R[(O^Y4]zu5N(62g(iL)ĖNtEv,\MeIbZI=:|* գL N0.I "'yÓ=ֈ8<3p9)/s}OGt­ʛA0$'\U.k^r] OrL%pQOZPe^:>&- y+\:g.Acm6 fu8?H_o6E)v?XWmtwdr;jЫ+ɪ,24aPg$Aq!̥k#0!& r7xc:r<RQcKEM3iऔ m|(WZp] }4!*pDyuʃD:ªM=#t&0 _ =X=(:ikR DjN=\w0*+fw.yG}>F $^ɕwvOߓե_U2QR3q͝0ϵ25zWk2k |\OF&?Pκd0596sE-`#\JVX'5}$Fu |N8_So-#w?]ᷟ[Pyn'@׻gkk,83 Vs ֯* n ЁVC\,,rDprk k8lۉK6u$B҂De-ЎZ r@׆E*t.qx/m*M5_zGhrx|o*+t/{U9ŇxaC3q]'Y~"ZLw063\# :9pO,L-w]NO9j'hݲz rZ V5{Nއa5XCb,qF{$!kFZcSj"9=1HFPNч;b(M|fT3AhK9eF0I"I(TmB{Dco@ҡ nI)Ia.J>FU:\KV6R~]n!ٯyS`^q6)xDI`EΑY<#}t!@%:}읊d?sW[6^ )HD#/hXX[:!+8"x"Ͽ A:GB4-G īm#6Bԗ(3~o8J'eKLAsRUށYQT LuѰb@ؠ|мSjMZ@ +kԕ9v[#N]\Ju>ے4ҞuzVcvO-{Eõ0P-5p83gw=2-r&xk!4bNA]z$dmP6+MizΚ:-jpb&9YB_n{ц?nrNJBO~Ǐ-Jd--螰f#wu֦f֦HOe4T`TEr+05ɭ%ZdB"ϒǾ褔u ǚ=/z>0l(!(Lu46kx$[Bl ?w|roaj\s`>B#V.H\-:V8{&4u=lr68:OqzBd0u~vur)Pudq!-.Ydz=?$8Ф\fnncpy%RTa!Iw1eYq$_nה6SD1 6U">6NWgQ}uFUP81V4jlIc\uI8s,j,mIVr 9`ky%I[Blr$UdX.MU<(rfiDZH[:Bʧ7, "_>R}FٟMe rk( 0SR;Z_`.F잾ԥЀk1:;1:atZ<ꤰt&G۴_&?d̚:|\]s]Ӟ*,`Y>G´sIgMmS9r(^WܻhN EsxE~Rnx\)_+9Euw樒Nb9!@}5K$ʞ<ΚD+VT'?aЈ_ܤIv"^&HRƬ"~>ʲʠ1t9Ǐ偆JTa@IPZ#z"Fh ;| @?PKUy`x}+\ӎbkty^4,Ck\U>ׁD..{O8)2H @l95ft{oijI}W(`ośmmKesCuݘ2wK /i ܿ!|iФM 2+A*`AqHӨ#bk57}{3z+ Sz=f޾)_nߠ@Ώ~)*ۯqo'ݞHf>>m/>@cWZ\ 6 ?Zj`Ns`w]Bf#UU⊔ܘ!@>82* *M"F! IC $('Hʇ1FBH,v3#챛6:U#jGm/KrR͉3e˶=+7i!yj2VF:Y̘ jccR;yē2IP-?VruzQ-ç| gE|oHT0QZ4 BEJ*n.B#⃣]nȾ:n&9R2FNtɿ3 Pd3~T}K(Qi8RJ?:ʼ#R>gI[2,S8N̙W#T)p|C k Zb͞OwK|/~Eq6ZhG'$QsT^Ь܏Ň hJ~~؝js;dGB~w J( E؏2JFa'sE:{#"&pڽkc-e*-j6F,!uQr#e2QEB.VL.&ae: ;*NȄ[d{@)!xO{5BmO{6Zg$drk冇a1z9Wk-rt+"Pu='3{CA1} u{%AO>k ܽo X+wvHyG!jxN=rJZ}6n ZɉCw>nӍX {.'&zh'Fn6OL3mSO=7֓On \:/jRYP9qza-/V0('LZEwn6ƱƱrSC^Ue{=crde67(Ah߃>KLnr־m+8}֌ h \圭Sp1u]Oǜnr'u N7bu`}R!ܗ)Uv vb{Xt󬃥-(Tn%^Qlw9mPp3_~%G+w5qT9fO).0 dTUhLaeᅵ| W'g''=#ܽ{St'zipV9dO~ZCsTg+-!QR1Ej UAYeU(PU:̀2bIPkL6{eQ'1Z][˦j&H6SH`@9֠4uFP{1DPRKĤ5ZudKtf$DsguI DVf unde .>¨fxYdTU눠fU ]R2&5 %\M ѝ{+fdi|5=C;؍hH-/]mnc0[/9{@tRJ5t8fVrwZlATy clV MR>];̻1+=!Ƒ`.BsM$)oO3.M'#]4{Tϕ`YXR??Rj)*`܁槠Fg]ߪ~Z,Eo)|0a͑yTRKP/+%2Q #E [W6"`btRU5}F3®f ˙>h( UdhMH:+{ 63Ui+TI )(] 0;LM'BEMRbYNX3Îvc*jޢvo;4GƯ{RMJ_Bgb51T/V3Jċgc~(ǹ-S0X U0 X"x8a\I}%"xoZqϛ{;h/7r%>Eٓ|ۦ6y Mۆi +4e.1s2~d5Nc' &ښ*2z:LqOJ2ez ,ZHG+Ύ#ܿ_gn3V/g{ i6撞-s0g!/7_n..o쌜h\ V~&!Fmu)# .- x sK^ͷ;IJR -~[0#Big.N6c4* .63u˩=>Ϣ'%{-Cvf,܃ό9)M@#KUJ)GO1eD e_u(9C&kq*VS0ѓEdLDSX0IFl 7JMh&1 Dȱd~IJSBRG՝&oP%#`<^i,m[8mJ j x+k%m]cFSolK(ejb+anhǪglyVnFR&h<(R#j-wI1Ї8c)z*̠˝,>#Dd,bgʰ(u"/W @0P&ls͠a>/VHQÐ".6 y D`BS%Ŏwܻ?,W1z8t!MNjgcZ2|v~D%-,z&)gC[9R&yL/}gfjm N Mi6ɧmOEYU}k 6LH4->ml/tpoRydqS7ufQm+՞[I=.tGOacD2ncp҉^O't~H({=Y@[XLЇ˃#lDNJew-]?1`I_x/C rQ lhfDeD@ WZ,A,1.rĻ.X_X:؎ Kʊ@ibeM;q&&JU2W&jDMl[ PK֗lclan TG0eR3rBGeF=oVzgI6V 8)cTJ:c 7 PڑS{^dRչZ*ؘX!RHeM:v|BPAZ1y*1)ڱk]!;mSV&hrXKQ9kd:%LH.*0d=)H1de] f堲j 1{ԁ&1Q_c&U(N@}&!O"jr%heAZK*;ʠAQ6ZCC*MɎE[*P%c Ud2'B,вLuLj)AMgcc8 ud1BK(ZO ׁUH=B^LU6x)MԾo5 \(18Rx;RmP*G,Plu5xWmzL`5^hI]xMg%"e [YԛTG&.E5 >[:n"۝>RlЏ 7 Q;z֓\5ԖH鼃Q wAu+lk'w %.uLąזW[=.5=0~Mgs N@ց3-;l]`lQ,IidIX\=)6xcVeaNjcyW&),tނ=L.PD0Ӎմ?!=j7קo}=˷!T:媼;?}pCz^ _>_L4|/קMO\6>0k V~ˬ~/CHƨ鱨JR%#sƐښå/?ݩ/O\]l1WK"*|.oqԗꉑyJctsnK0`%foݷl1Žte(r[Vr^wv^%H uRW7ר7)}}Rj]es+bL ZgU8 >t}H `="ZgfćᲧѾ YZ#I z6eNzIR7Kgiu DzoБQYAGpHu$iH93}_oyEڛ?=,AZy6@9g/Wx}$9nx=ۚg 2$`h9j],67͍3',M7J*xFuV+hy&dWeIE6t QPqdWѣS!ˑ$=b&:R? S-XC먀` bfSUYyeMM5D` SȦϠz915ܚv5ޒsk2ܝra5-yYɂ9fȠz4oI<E[#~V<舒sqݭZfy_>) ]kq`hQi, Smd]έŠWa.gь i$/`'H>Rkyy.>ܮOcg׶W taSE"em9 8+~$6I[TʮfPU|^2/^&-Ϯo#X&S9>6R)6JJL"AundgKZb. j^MѮ:^s>7S^ϭ/.n(ϖ&~3qyUVgܷSTf֜wmm%W~Jb_3Hdނ`lI(ȿjJ([HsCO,Z/Q˒dgkv~}89:_B=ulEFb2үJUx\z[b tL>%|=] KV7Vy $ɗƆn?TdZ h%%"겓Ι->5i"Ы d@K68_,eVF|l+LUGUдZR5ٜ}p()E\|T–9!(ہصA2XA&8䢛R/^ Gl#'u3ci>Wg]rլ|20U^]D-#6!<6DSEbDy!9Lbn_SƅƏ̋#%&֠Շ-j.w1EgC4rN94,$K$z9e['d !1f9x7;N۲| Z\1YXE@"Pu@ݞm >NNc"}})yCRPQ'D%+txm5C.y![YZ]bBjfu}Iv~d|z-دǛ͐AmH-j HbL9 a|n=|6Ё%pbtcU0`.\Tũ\< b<~Uw+E[98+1{N4$j:'td+2DQ7S0:ifpGWxLrlO *hMPPlљ|1x)ȏS !"?ֺo(sԡoLGŅBPB}}\Bk VCJYr=/FїTŶ ʗ| W: Wq6aH 9YZur/̌XHCa(kT3}HFLЕyg&\g"7wdSTB@Lk]|O6OK G_ WP0#'Ur?N PgC¡rP(B%;H:}q豗 R|(LZk&$HhN T#C0 CO28"":[#noR J#V21%m||h_\N˜:M|(P8raHEo*r&]aOzZ/OA(_Pt7KKA'BnSaE-AKZU -OX#l>2|肈$s1~=-=?h7W -%jy@j 6ľg)ZK6r"d]IS1`lZ Vv;c dp|-ŃNd&f$I_TU c>R'oyW^''&@}T޺" wD&tuHC,8 xn }=*pW%:ռeuc 'Zrۦ 0n|8̶e^a".>z+tl̒vG133xP 1y^e&4RVyʹV.u`j ޕCh |SUV+cz-N[|mƵEȫA,/K/SI}kLhܖEvSE/:hXM}SEe5ehŏ!7.E.P7Ef/P&ސZs-WwՌr.(UtCI '?_;U owmf6/3$~Ǒ~s߷mU\ؒhǨJ]:goWDQktv`C[,b~7=TT>KYk;yh8X`x =F-nφ=dQYHRt/ƦГs}F'?Q`A7`DؘQꐕEu2D ;;, C9B tq;hB.*BSA@g4)<xbC }`%7u7]Ќ9WU4(ǦИ,=Ӈ>ͮO)AiU·oCSPl${EV56`,zꖥ7B^͛p7WC+%-w B1H`/3b:5~sơ?|}^ĥQ!M1F{!-hJ=ɥ)&=SQ/hN7~/9ӷDŕ0.BÀBk Zx]Ugy%<ćVFćK{)E$$%xDe+L)*ZUi?1֣x>)̑qv8O8wnMy UvTe+d\xp/Db9߸}G.fR"*׉lA=a]*[g"!5LAX,=8=Jj6] #9DԢ-tkl~ڒkx::^,g$f9 ǻahr޷R`"̅A5)H#j,Ѧ46~y'/H㪚6zc^«ǃͽ.g9Q Oc1X#2ﲌ(ia 1jg> z+|?_$Eݔߋ^$$Ï6YF;ظ|$IpL3>(l6&#q?$A.]B1Gߎ5ybxJBv 0}ȓkEEdt"O[$C8 Fvo&}#B%-~nZp%F6"{-DcnV"{[])6_Z%bO.뺮l8\߶9e&#^ȯ\;U*Y彲_.W_?YtCz/KK-fW{;y y[m]ڛLr>}?C'/Wsq$?ݏ]N>htȿ9.Zmr 6"uLz)&գOa3XS=7jD%x3ƫ!`(?YԵ9K78;ZC?NZ2T3:t{;7wc #4k_?)@fM6lR: JT/&5 +F)fZ);mqC"Q/·[۳ ,Xhg 鿟E\IHlm1fwśZ~Vk[01K̙;;w)?NMkT5 i.9"RDޕq$e/X x98~q@)1vzHCR%rWUՌrmlJ($$ƽ~]vata';}b߀eJZwN+à~p? `8( [bЏ<읺? ne }8]#V*L$9؃bճ>;?B_⨓> L ^đ%qчxQ.yQ8+}4f?6:\#EAƽUi)s wK_l3% ^tWʂ2]g҉㻄{g:z©*G.9D\́r-(DțCF|+Hp Jś4Ja:F_V4o n_dyQAb<E*5>^5à#D[d@FhKSʂ 6~-y(8;~(U\Y,q~]hd;kd;kd;kd3@4ߛW3g%X;ł;@<"9l j-enޕ_\釕iGŻr3C\3 9Ozۤ쳘\J&V'4|M;dMU-I{i;,/}3fhIz/{ؚ3gf$yY0ż8a>a5"Hѐ"ٙ 7JIl9Wjw{S0k X6y4tl8 oajbZ^ҏC.B>S4,XKxg0X LeR8`(M{t>&lv)h@pH e!̩-Qy/[8N]Ɂ!8(^yv{lIsexǜ];,nA\ &d=sg`O~ʌmj#UHKA,whtFШYs ˋ'ͳ-%gV:L}kg־[eì#3Qo_˓@B""w8¢ىN gyM I2axd)~o5?A^N:}Y@|8kSo 6'n0F"6Ɖ;pϴRk9/AGaBD)((JcĕxR,x^m%'^\.r7N{-5 Q_0R֧o \0 '1Z9;+4%J0k4@(Ï2A\,!A% Ȑ3#BJ o$>_o)J qY^LkE`D>s-]v3yN6& kE\D#'VҌrZT~?>SEMY 2=O\V~d%jJ_Z%s-±@ es; ;F!Fv_ݣܑZ+Փ#\L0ͩ[.uHk:Z4 ^Ǫ"b~:g糔QۖM˱.QE c t JF<O惸9ڂJݍ JP.;t?]ϥ\qqϗ{돹VeX4hK\ЬdncVͱMUJRqO9/2KF Ss^dlk gb<!5 &-W+sY_bE%keO2EwGBO:{jBYF3vdJƨa8YS/Msa̅HʕYa%x |j#˘5|8R "9xRG7]X |dpNYoeH2cא&Q&;5Tk6ɑӦy&1ʐ)vEMxV"Ėd o{xIQC[]b[F$[+?A4;»?"dMbeK`ĭQ홿ۗFOLM`4.&f`:LA\30((4PkiZy׳:->ۛyѫcadc1N~>W#MtL=6\7Nǝqd]\Մ\1R~lw+Z J1hS:_<Ǘ?Eȹ"G! TJiQc_.d,/^sSŋZo[Xvk,X -5%I%.dj Du=k 8 ܋}?L#NBk[Gs> T{}4K`DO:83<Շ=#4MXD_ ˏ ONƞLeeYd.{Rg7a,W +,Ny~UTD9ۈ=e[S N>vn9 rF֭yۚUtMtM-)]aY/%ia[ɼii~ azqb]ƟǓkW-V9bGBɧ6b:9ĩW@Ry:ōvcT+l)2zAL/~mm5WVFidD-IRޝy10%(A-פLp^+ a.!Dx7q% 3q0rIq0b$69zvT}%48rb@"1 uXE LJx2pն\HC'B8, (ɂ& V/@B؁n/w" D)~@|qǂ b&ZJ/% ٹ Dj:1TNqMtSc2*;I~2ؘF q ω_!<}~r\5y>l"?nצ Bz}u.1w&+܊omFd}f9J0Ҝj"̤ҭ/6ZŦrJav*2Q|tLE0AdFh\&t=]0 ;z{ba O6FgmBJ#rdQ )Zjʽ#Uҩ6N(9r0ODA8 XZsUw Q֞G?Tbs $ k\YDXKp}[S~k:[FȾ$e`:-)2 tTQ䷮؟xڝ̊dC45e!JYr݋V]@UnZ{Q* {rFԾY~8m' 9nbI |X1r0̊M@}xܹ"ɿ"Kpi,$ ns 6/ $m$9I0-ɦ^6q&R?Ū"륓Qq=t>׉ :m؎ z Leh,I͒!Q(#0$M 駔3LydLQIɴ q5wܑ!f5?"푈,S]K$n9Ojߟ~B{N8؄l -;L]uYw7}~g}QbВb;V9HN%.:lFrH,MT8;?Mˣ,YL##<<κ8xFOG 8 ~u_>|k<=s$Kpqg'J /Ωn @R4lL׍O'>խ@<cVBfbYDjbjbYC]jbmƋ|z'8ejc7xvS/ ꀒ̺\'eD@G&}OE).əDBP"էg_/P@q?3P7bY'T$DPMN ?jp8Juʝ}E]~$VS8^S@_>^]_G"Y yY8%وf5CT2sˑsCջtnDJL6΢s Ei2lYU$ǡZ{a,yé1sWsU!"J6Ւ΄Z#rͯh!,$㴀".6 FI ^s1•6EAdU1>URSMS~$C{q{# :V^m4,Sތi86iҤXR8F|ǐ gɁm.Q:,Y> zDG0yF=F cv>Cf 0c¤9\1 W@ qm9TP$]E+x"VbԄ-h8^뇲ĄX]pct=\uy-m=52#)9PW7%@QTXS@yW մ5oz* Dϴ[ `p.MovOi1 mqP̏uN3IѺ#V,g2^H+fMťtZIkL{V*Kt/1I */٘aF=L]hy4܊ jZ9' j W*I51^xY",xt"#$#s&?"-Ay/T gTD@!KKF9 [SVm2B&aQ);R~֌:Ҿ(ֻL1WXQL8R>s%P۪*pWLT8JL%=12Tj*N݇}{Wer|Qvz>z[]5]~G]1ƀ?|-|3YÏ&EiML84aE3F?*lu=K+fsn\~h gk;jx^D =JX3Θy%~{ϊz^$Hx}Fg1].ɻbuo3Wo*4!m2{czjoJ>,%U~e[\ʕ1&'mv4t,n]yйFs;=Ohfx868G; N}ns3mߣ,l C>L^ 8qqQp.$OMTHeG B c)p"&W`gU윋x|u/-!q[+V6n0zieӌIi^ w :{}AO#:5aD oU$[\5cϦ?!hI?DDt&A pnvGInٌsFk p2T͘[1EЂg.~>g. ;>s1)=@]4_GwKXVV <=!]$ӉŖƟ7f9 Sڣo)|'![]XT2J{Q׈d>{9iܛj9.H:sP-RGVNjծ>j%G?[ѿ }9^u7eY `tkش7rĤJ TtqF2vj5_:QRN/㦝&VERz*_2Hrz7 ݀#NT[nj&uyn02A2Tn=a1X$Lnpd$Rh[ǔ0@yaHȻTLC2tu5` n @`z40M+>9LX&ez- JÕ?i.~j5lf1#p3x}#9(<TiSé.$J^YbS$! kۆNXӌpyjYu`M%5V|Wn0zlQ-[OphPYQ30ǫGk Jr0"CA}ቆ$P Tz3*_> 6$!,.PkN_y$ k\=ԌH G>Rmn pZy6DH݂Z$w>HklTM'7Q(j'4,zntzq6s4+5]{2kT7&]ytc|3 un3 ,[ t}лp? #f1S|b>.ņUF1:82uwtҳh.T=-&!n|D֕9M1w ݺgj>$  fb7+'~nVyOuynV9.^7髛m5~>)lZ{hzBH tv~ IמjqӐۆV]zFaKc>=MY.YmY%5fJM$xRܹՑ׉:S>iP: S? upKI}XW:F-2fY',c7b4$z@i d@E`  ]-$InDZW[ݿo :kR_~:RĻ)#RӢ$ )\Gb;VQ3ɒy&VA`8Z8L7xM+pC/1@S66X3󘃐8-ϣ>xᛣ-^U|ޑQ-GDc?DC|.W$!tFp BSQIRIuťC0F|~|=Xcwtuoj6ܓuAqu4Ib\U^܉g!dlLXZ=U"؂V1JV+QQhDb'/&yjb=T1WQYAA xe) gP_aIf \eKNJC=߀ZdLȱP.?mjdn:y~Q-~WnNbHZDvjxۻ߮LU뤘at? ~~9Z_wԘN?_t;^n~EE:{1vs?Ze('E"d#yjVIx]k'ɞbMTɸ\BW*%(KM%" 9s6kR^wŋx]QM |S"$&} B$.&ߌ"-/nӧ9>3J=؝Ӏ:.Gnq9;EݑP+7+XZrc昫ii"ԞyZ8eWivfrQ#~CJ~&8,&7 ;wVsٰ#^rUA݌#.y>P DQ;8)qj,gdw۽ v| Z{VO7mds_L\zТ:kP{n&\lQsͺ#͏|u{8puSbYebdZFh:2">Bz1-WA56Fv{PclW6iFdj~Q=~>MZjM"mވ[/dP?z#o+1o%YhPrIV0K"' ]C/ Vim1d_9HLhDN@%5U;HBO<d?s _bpfub7A:5T$4ZFiF|\WڬEb~L;V!qL9/% S(`=9kI$^+ pFzhF".> .-ZK0AiOXu~ͫ&$a# fHBBXRxԋ4ۦuB Xboݍ1&fY&c-1}Ii\܈M^խ]a]Ģ0/[l7f?('@߯{+5=I?MO~nuȁ*4M3(6LaXiQSd{G ^Deنc: s:igmlBO kzנ_Q!4585k?w7ۚfPs,]0kcIߣHQ}%e{|Q tBτ&,'#0]̼ *A9֣꓇=˳!LVbvM)ٜ4f>7y  pӺP$~XU?tG2,9'%djU6l192.168.126.11:10357: read: connection reset by peer" start-of-body= Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.254848 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:46140->192.168.126.11:10357: read: connection reset by peer" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.254914 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.255068 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.256222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.256274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.256285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.257010 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"4557fcde61f57116f59d3965fc85f08395545a7008311738e24afc920c4fbde2"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.257169 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://4557fcde61f57116f59d3965fc85f08395545a7008311738e24afc920c4fbde2" gracePeriod=30 Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.425109 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:13Z is after 2026-02-23T05:33:13Z Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.489228 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.490867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.490901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.490912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.491462 4926 scope.go:117] "RemoveContainer" containerID="0bb2797eb4e103e38988a8ca21a8dbc1acd77e1a462c8c15d187a78d48b0fdac" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.652856 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.653716 4926 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="4557fcde61f57116f59d3965fc85f08395545a7008311738e24afc920c4fbde2" exitCode=255 Mar 12 18:03:13 crc kubenswrapper[4926]: I0312 18:03:13.653771 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"4557fcde61f57116f59d3965fc85f08395545a7008311738e24afc920c4fbde2"} Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.425272 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:14Z is after 2026-02-23T05:33:13Z Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.659418 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.661169 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97"} Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.661334 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.662148 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.662173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.662181 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.663551 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.663909 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8"} Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.663968 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.664699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.664743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:14 crc kubenswrapper[4926]: I0312 18:03:14.664760 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:14 crc kubenswrapper[4926]: E0312 18:03:14.935532 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:14Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c2a06703cc5e3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,LastTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.341936 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:15 crc kubenswrapper[4926]: E0312 18:03:15.342687 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:15Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.343257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.343301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.343316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.343342 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 18:03:15 crc kubenswrapper[4926]: E0312 18:03:15.345798 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:15Z is after 2026-02-23T05:33:13Z" node="crc" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.425697 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:15Z is after 2026-02-23T05:33:13Z Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.668655 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.669594 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.671742 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97" exitCode=255 Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.671821 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97"} Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.671904 4926 scope.go:117] "RemoveContainer" containerID="0bb2797eb4e103e38988a8ca21a8dbc1acd77e1a462c8c15d187a78d48b0fdac" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.671940 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.672112 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.673645 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.673685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.673715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.674102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.674165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.674189 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:15 crc kubenswrapper[4926]: I0312 18:03:15.675289 4926 scope.go:117] "RemoveContainer" containerID="237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97" Mar 12 18:03:15 crc kubenswrapper[4926]: E0312 18:03:15.675697 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:03:16 crc kubenswrapper[4926]: I0312 18:03:16.426226 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:16Z is after 2026-02-23T05:33:13Z Mar 12 18:03:16 crc kubenswrapper[4926]: I0312 18:03:16.675854 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 18:03:17 crc kubenswrapper[4926]: I0312 18:03:17.428061 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:17Z is after 2026-02-23T05:33:13Z Mar 12 18:03:17 crc kubenswrapper[4926]: I0312 18:03:17.572563 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:17 crc kubenswrapper[4926]: I0312 18:03:17.572743 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:17 crc kubenswrapper[4926]: I0312 18:03:17.574185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:17 crc kubenswrapper[4926]: I0312 18:03:17.574215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:17 crc kubenswrapper[4926]: I0312 18:03:17.574225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:18 crc kubenswrapper[4926]: I0312 18:03:18.426128 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:18Z is after 2026-02-23T05:33:13Z Mar 12 18:03:19 crc kubenswrapper[4926]: I0312 18:03:19.427547 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:19Z is after 2026-02-23T05:33:13Z Mar 12 18:03:19 crc kubenswrapper[4926]: I0312 18:03:19.702285 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 12 18:03:19 crc kubenswrapper[4926]: E0312 18:03:19.708767 4926 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:19Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 18:03:19 crc kubenswrapper[4926]: E0312 18:03:19.710482 4926 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 12 18:03:20 crc kubenswrapper[4926]: I0312 18:03:20.424741 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:20Z is after 2026-02-23T05:33:13Z Mar 12 18:03:20 crc kubenswrapper[4926]: E0312 18:03:20.563991 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.229553 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.229857 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.231394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.231467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.231486 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.232257 4926 scope.go:117] "RemoveContainer" containerID="237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97" Mar 12 18:03:21 crc kubenswrapper[4926]: E0312 18:03:21.232709 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.279047 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.279248 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.280365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.280593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.280736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.337857 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.427202 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:21Z is after 2026-02-23T05:33:13Z Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.691956 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.693490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.693550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.693567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:21 crc kubenswrapper[4926]: I0312 18:03:21.694366 4926 scope.go:117] "RemoveContainer" containerID="237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97" Mar 12 18:03:21 crc kubenswrapper[4926]: E0312 18:03:21.694656 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:03:22 crc kubenswrapper[4926]: I0312 18:03:22.346353 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:22 crc kubenswrapper[4926]: I0312 18:03:22.348285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:22 crc kubenswrapper[4926]: I0312 18:03:22.348348 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:22 crc kubenswrapper[4926]: I0312 18:03:22.348372 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:22 crc kubenswrapper[4926]: I0312 18:03:22.348418 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 18:03:22 crc kubenswrapper[4926]: E0312 18:03:22.348762 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:22Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 12 18:03:22 crc kubenswrapper[4926]: E0312 18:03:22.353564 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:22Z is after 2026-02-23T05:33:13Z" node="crc" Mar 12 18:03:22 crc kubenswrapper[4926]: I0312 18:03:22.427168 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:22Z is after 2026-02-23T05:33:13Z Mar 12 18:03:22 crc kubenswrapper[4926]: W0312 18:03:22.794830 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:22Z is after 2026-02-23T05:33:13Z Mar 12 18:03:22 crc kubenswrapper[4926]: E0312 18:03:22.794909 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 18:03:23 crc kubenswrapper[4926]: I0312 18:03:23.424999 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:23Z is after 2026-02-23T05:33:13Z Mar 12 18:03:24 crc kubenswrapper[4926]: I0312 18:03:24.280042 4926 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 18:03:24 crc kubenswrapper[4926]: I0312 18:03:24.280192 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 18:03:24 crc kubenswrapper[4926]: I0312 18:03:24.427842 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:24Z is after 2026-02-23T05:33:13Z Mar 12 18:03:24 crc kubenswrapper[4926]: W0312 18:03:24.875317 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:24Z is after 2026-02-23T05:33:13Z Mar 12 18:03:24 crc kubenswrapper[4926]: E0312 18:03:24.875395 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 18:03:24 crc kubenswrapper[4926]: E0312 18:03:24.939541 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:24Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c2a06703cc5e3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,LastTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:25 crc kubenswrapper[4926]: I0312 18:03:25.425005 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:25Z is after 2026-02-23T05:33:13Z Mar 12 18:03:26 crc kubenswrapper[4926]: I0312 18:03:26.424818 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:26Z is after 2026-02-23T05:33:13Z Mar 12 18:03:27 crc kubenswrapper[4926]: I0312 18:03:27.371025 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 18:03:27 crc kubenswrapper[4926]: I0312 18:03:27.371224 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:27 crc kubenswrapper[4926]: I0312 18:03:27.372207 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:27 crc kubenswrapper[4926]: I0312 18:03:27.372243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:27 crc kubenswrapper[4926]: I0312 18:03:27.372252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:27 crc kubenswrapper[4926]: I0312 18:03:27.425512 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:27Z is after 2026-02-23T05:33:13Z Mar 12 18:03:28 crc kubenswrapper[4926]: I0312 18:03:28.425511 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:28Z is after 2026-02-23T05:33:13Z Mar 12 18:03:29 crc kubenswrapper[4926]: I0312 18:03:29.353749 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:29 crc kubenswrapper[4926]: E0312 18:03:29.353855 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:29Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 12 18:03:29 crc kubenswrapper[4926]: I0312 18:03:29.354724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:29 crc kubenswrapper[4926]: I0312 18:03:29.354757 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:29 crc kubenswrapper[4926]: I0312 18:03:29.354767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:29 crc kubenswrapper[4926]: I0312 18:03:29.354789 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 18:03:29 crc kubenswrapper[4926]: E0312 18:03:29.357668 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:29Z is after 2026-02-23T05:33:13Z" node="crc" Mar 12 18:03:29 crc kubenswrapper[4926]: I0312 18:03:29.426603 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:29Z is after 2026-02-23T05:33:13Z Mar 12 18:03:30 crc kubenswrapper[4926]: I0312 18:03:30.426785 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:30Z is after 2026-02-23T05:33:13Z Mar 12 18:03:30 crc kubenswrapper[4926]: E0312 18:03:30.564137 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 18:03:31 crc kubenswrapper[4926]: I0312 18:03:31.425750 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:31Z is after 2026-02-23T05:33:13Z Mar 12 18:03:32 crc kubenswrapper[4926]: W0312 18:03:32.395234 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:32Z is after 2026-02-23T05:33:13Z Mar 12 18:03:32 crc kubenswrapper[4926]: E0312 18:03:32.395311 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:32Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 18:03:32 crc kubenswrapper[4926]: I0312 18:03:32.427936 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:32Z is after 2026-02-23T05:33:13Z Mar 12 18:03:33 crc kubenswrapper[4926]: I0312 18:03:33.425124 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:33Z is after 2026-02-23T05:33:13Z Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.279846 4926 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.279906 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.425061 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:34Z is after 2026-02-23T05:33:13Z Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.489351 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.490587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.490619 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.490630 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:34 crc kubenswrapper[4926]: I0312 18:03:34.491152 4926 scope.go:117] "RemoveContainer" containerID="237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97" Mar 12 18:03:34 crc kubenswrapper[4926]: E0312 18:03:34.491328 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:03:34 crc kubenswrapper[4926]: E0312 18:03:34.945321 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:34Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c2a06703cc5e3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,LastTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:35 crc kubenswrapper[4926]: W0312 18:03:35.402972 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:35Z is after 2026-02-23T05:33:13Z Mar 12 18:03:35 crc kubenswrapper[4926]: E0312 18:03:35.403095 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:35Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 18:03:35 crc kubenswrapper[4926]: I0312 18:03:35.425190 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:35Z is after 2026-02-23T05:33:13Z Mar 12 18:03:36 crc kubenswrapper[4926]: I0312 18:03:36.358166 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:36 crc kubenswrapper[4926]: I0312 18:03:36.359581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:36 crc kubenswrapper[4926]: I0312 18:03:36.359644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:36 crc kubenswrapper[4926]: I0312 18:03:36.359669 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:36 crc kubenswrapper[4926]: I0312 18:03:36.359711 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 18:03:36 crc kubenswrapper[4926]: E0312 18:03:36.361958 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:36Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 12 18:03:36 crc kubenswrapper[4926]: E0312 18:03:36.364407 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:36Z is after 2026-02-23T05:33:13Z" node="crc" Mar 12 18:03:36 crc kubenswrapper[4926]: I0312 18:03:36.424625 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:36Z is after 2026-02-23T05:33:13Z Mar 12 18:03:37 crc kubenswrapper[4926]: I0312 18:03:37.424899 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:37Z is after 2026-02-23T05:33:13Z Mar 12 18:03:38 crc kubenswrapper[4926]: I0312 18:03:38.426981 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:38Z is after 2026-02-23T05:33:13Z Mar 12 18:03:39 crc kubenswrapper[4926]: I0312 18:03:39.425545 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:39Z is after 2026-02-23T05:33:13Z Mar 12 18:03:40 crc kubenswrapper[4926]: I0312 18:03:40.426184 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:03:40Z is after 2026-02-23T05:33:13Z Mar 12 18:03:40 crc kubenswrapper[4926]: E0312 18:03:40.564287 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 18:03:41 crc kubenswrapper[4926]: I0312 18:03:41.429325 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:42 crc kubenswrapper[4926]: I0312 18:03:42.428888 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:43 crc kubenswrapper[4926]: I0312 18:03:43.364725 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:43 crc kubenswrapper[4926]: I0312 18:03:43.367503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:43 crc kubenswrapper[4926]: I0312 18:03:43.367619 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:43 crc kubenswrapper[4926]: I0312 18:03:43.367642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:43 crc kubenswrapper[4926]: I0312 18:03:43.367690 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 18:03:43 crc kubenswrapper[4926]: E0312 18:03:43.369321 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 18:03:43 crc kubenswrapper[4926]: E0312 18:03:43.370149 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 18:03:43 crc kubenswrapper[4926]: I0312 18:03:43.431671 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.278542 4926 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.278686 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.278773 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.278968 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.280242 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.280291 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.280308 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.280949 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.281099 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8" gracePeriod=30 Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.428653 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.750201 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.751370 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.751811 4926 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8" exitCode=255 Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.751854 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8"} Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.751880 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"405f59d0da6c9a3663ed746f08f9d5c2d94818971dbc0ce0373690c731b5afae"} Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.751898 4926 scope.go:117] "RemoveContainer" containerID="4557fcde61f57116f59d3965fc85f08395545a7008311738e24afc920c4fbde2" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.752003 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.752933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.752988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:44 crc kubenswrapper[4926]: I0312 18:03:44.753005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.949381 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06703cc5e3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,LastTimestamp:2026-03-12 18:02:40.418334179 +0000 UTC m=+0.786960552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.954571 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.960461 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.964034 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e8176 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,LastTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.967558 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a067856139d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.554210205 +0000 UTC m=+0.922836558,LastTimestamp:2026-03-12 18:02:40.554210205 +0000 UTC m=+0.922836558,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.971675 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e238d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.591277422 +0000 UTC m=+0.959903755,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.975104 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e5dfe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.591295923 +0000 UTC m=+0.959922256,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.980649 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e8176\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e8176 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,LastTimestamp:2026-03-12 18:02:40.591325114 +0000 UTC m=+0.959951447,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.986428 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e238d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.592684351 +0000 UTC m=+0.961310684,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.989840 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e5dfe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.592703172 +0000 UTC m=+0.961329495,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.993327 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e8176\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e8176 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,LastTimestamp:2026-03-12 18:02:40.592713192 +0000 UTC m=+0.961339525,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:44 crc kubenswrapper[4926]: E0312 18:03:44.998549 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e238d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.593225781 +0000 UTC m=+0.961852114,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.002382 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e5dfe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.593236761 +0000 UTC m=+0.961863094,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.005735 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e8176\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e8176 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,LastTimestamp:2026-03-12 18:02:40.593247111 +0000 UTC m=+0.961873444,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.009822 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e238d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.593868664 +0000 UTC m=+0.962495007,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.014508 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e5dfe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.593885395 +0000 UTC m=+0.962511728,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.023359 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e8176\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e8176 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,LastTimestamp:2026-03-12 18:02:40.593896115 +0000 UTC m=+0.962522448,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.029380 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e238d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.594219377 +0000 UTC m=+0.962845710,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.035891 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e5dfe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.594248538 +0000 UTC m=+0.962874871,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.042105 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e8176\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e8176 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,LastTimestamp:2026-03-12 18:02:40.59428994 +0000 UTC m=+0.962916273,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.046295 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e238d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.594938093 +0000 UTC m=+0.963564426,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.050529 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e5dfe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.594956754 +0000 UTC m=+0.963583087,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.054501 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e8176\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e8176 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474022262 +0000 UTC m=+0.842648595,LastTimestamp:2026-03-12 18:02:40.594966864 +0000 UTC m=+0.963593187,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.058127 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e238d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e238d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.473998221 +0000 UTC m=+0.842624554,LastTimestamp:2026-03-12 18:02:40.595792374 +0000 UTC m=+0.964418707,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.062940 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c2a06738e5dfe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c2a06738e5dfe default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.474013182 +0000 UTC m=+0.842639515,LastTimestamp:2026-03-12 18:02:40.595807374 +0000 UTC m=+0.964433707,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.068804 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a0691e1fd66 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:40.982809958 +0000 UTC m=+1.351436291,LastTimestamp:2026-03-12 18:02:40.982809958 +0000 UTC m=+1.351436291,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.073508 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c2a0692e9e595 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.000105365 +0000 UTC m=+1.368731758,LastTimestamp:2026-03-12 18:02:41.000105365 +0000 UTC m=+1.368731758,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.078421 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0692fefd8b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.001487755 +0000 UTC m=+1.370114088,LastTimestamp:2026-03-12 18:02:41.001487755 +0000 UTC m=+1.370114088,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.083508 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a06949e358c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.028699532 +0000 UTC m=+1.397325865,LastTimestamp:2026-03-12 18:02:41.028699532 +0000 UTC m=+1.397325865,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.088126 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06961efa2f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.053915695 +0000 UTC m=+1.422542028,LastTimestamp:2026-03-12 18:02:41.053915695 +0000 UTC m=+1.422542028,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.092971 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a06b3bba266 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.550721638 +0000 UTC m=+1.919347971,LastTimestamp:2026-03-12 18:02:41.550721638 +0000 UTC m=+1.919347971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.098840 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06b3d4273a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.552328506 +0000 UTC m=+1.920954839,LastTimestamp:2026-03-12 18:02:41.552328506 +0000 UTC m=+1.920954839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.102790 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a06b440585d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.559418973 +0000 UTC m=+1.928045306,LastTimestamp:2026-03-12 18:02:41.559418973 +0000 UTC m=+1.928045306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.107469 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c2a06b4c723a1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.568252833 +0000 UTC m=+1.936879166,LastTimestamp:2026-03-12 18:02:41.568252833 +0000 UTC m=+1.936879166,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.109802 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06b4d11a59 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.568905817 +0000 UTC m=+1.937532160,LastTimestamp:2026-03-12 18:02:41.568905817 +0000 UTC m=+1.937532160,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.111828 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a06b4d16a5a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.568926298 +0000 UTC m=+1.937552641,LastTimestamp:2026-03-12 18:02:41.568926298 +0000 UTC m=+1.937552641,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.114547 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a06b4db8c64 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.569590372 +0000 UTC m=+1.938216725,LastTimestamp:2026-03-12 18:02:41.569590372 +0000 UTC m=+1.938216725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.116123 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06b4e37904 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.5701097 +0000 UTC m=+1.938736043,LastTimestamp:2026-03-12 18:02:41.5701097 +0000 UTC m=+1.938736043,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.118152 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a06b5e85e1e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.58720771 +0000 UTC m=+1.955834043,LastTimestamp:2026-03-12 18:02:41.58720771 +0000 UTC m=+1.955834043,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.120812 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c2a06b5fcc745 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.588545349 +0000 UTC m=+1.957171682,LastTimestamp:2026-03-12 18:02:41.588545349 +0000 UTC m=+1.957171682,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.121988 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a06b6035769 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.588975465 +0000 UTC m=+1.957601788,LastTimestamp:2026-03-12 18:02:41.588975465 +0000 UTC m=+1.957601788,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.126137 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06c74187ff openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.878263807 +0000 UTC m=+2.246890180,LastTimestamp:2026-03-12 18:02:41.878263807 +0000 UTC m=+2.246890180,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.130185 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06c7ece494 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.889494164 +0000 UTC m=+2.258120527,LastTimestamp:2026-03-12 18:02:41.889494164 +0000 UTC m=+2.258120527,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.134269 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06c801c6ac openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.890862764 +0000 UTC m=+2.259489137,LastTimestamp:2026-03-12 18:02:41.890862764 +0000 UTC m=+2.259489137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.138754 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06d60c379c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.12642806 +0000 UTC m=+2.495054433,LastTimestamp:2026-03-12 18:02:42.12642806 +0000 UTC m=+2.495054433,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.143148 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06d6eec9e8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.141276648 +0000 UTC m=+2.509903021,LastTimestamp:2026-03-12 18:02:42.141276648 +0000 UTC m=+2.509903021,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.146760 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06d701b48b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.142516363 +0000 UTC m=+2.511142736,LastTimestamp:2026-03-12 18:02:42.142516363 +0000 UTC m=+2.511142736,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.149937 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06e1a8997d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.321226109 +0000 UTC m=+2.689852452,LastTimestamp:2026-03-12 18:02:42.321226109 +0000 UTC m=+2.689852452,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.153201 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06e240ddf6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.33120511 +0000 UTC m=+2.699831463,LastTimestamp:2026-03-12 18:02:42.33120511 +0000 UTC m=+2.699831463,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.157887 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a06ecb30c3d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.506460221 +0000 UTC m=+2.875086554,LastTimestamp:2026-03-12 18:02:42.506460221 +0000 UTC m=+2.875086554,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.162595 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a06eced1d3c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.51026566 +0000 UTC m=+2.878891993,LastTimestamp:2026-03-12 18:02:42.51026566 +0000 UTC m=+2.878891993,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.166548 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c2a06ed2f3e2a openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.514599466 +0000 UTC m=+2.883225829,LastTimestamp:2026-03-12 18:02:42.514599466 +0000 UTC m=+2.883225829,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.171550 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a06ed4c2a93 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.516494995 +0000 UTC m=+2.885121328,LastTimestamp:2026-03-12 18:02:42.516494995 +0000 UTC m=+2.885121328,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.175308 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a06f9cb7081 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.726162561 +0000 UTC m=+3.094788894,LastTimestamp:2026-03-12 18:02:42.726162561 +0000 UTC m=+3.094788894,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.179043 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a06f9e98f7b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.728136571 +0000 UTC m=+3.096762904,LastTimestamp:2026-03-12 18:02:42.728136571 +0000 UTC m=+3.096762904,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.182254 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a06f9ea0fc5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.728169413 +0000 UTC m=+3.096795756,LastTimestamp:2026-03-12 18:02:42.728169413 +0000 UTC m=+3.096795756,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.185752 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c2a06fa34bc77 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.733063287 +0000 UTC m=+3.101689620,LastTimestamp:2026-03-12 18:02:42.733063287 +0000 UTC m=+3.101689620,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.191067 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a06fad6380f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.743646223 +0000 UTC m=+3.112272556,LastTimestamp:2026-03-12 18:02:42.743646223 +0000 UTC m=+3.112272556,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.194582 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a06fae60544 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.744681796 +0000 UTC m=+3.113308129,LastTimestamp:2026-03-12 18:02:42.744681796 +0000 UTC m=+3.113308129,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.198814 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a06fafcfc5e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.746186846 +0000 UTC m=+3.114813199,LastTimestamp:2026-03-12 18:02:42.746186846 +0000 UTC m=+3.114813199,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.202292 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a06fb140298 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.747695768 +0000 UTC m=+3.116322111,LastTimestamp:2026-03-12 18:02:42.747695768 +0000 UTC m=+3.116322111,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.206891 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a06fb8e1d27 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.755697959 +0000 UTC m=+3.124324282,LastTimestamp:2026-03-12 18:02:42.755697959 +0000 UTC m=+3.124324282,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.210329 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c2a06fbc62132 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.75936901 +0000 UTC m=+3.127995343,LastTimestamp:2026-03-12 18:02:42.75936901 +0000 UTC m=+3.127995343,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.212276 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a0705d2c206 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.927968774 +0000 UTC m=+3.296595107,LastTimestamp:2026-03-12 18:02:42.927968774 +0000 UTC m=+3.296595107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.214022 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a0705d36709 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.928011017 +0000 UTC m=+3.296637350,LastTimestamp:2026-03-12 18:02:42.928011017 +0000 UTC m=+3.296637350,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.216433 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a070688522a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.93986769 +0000 UTC m=+3.308494023,LastTimestamp:2026-03-12 18:02:42.93986769 +0000 UTC m=+3.308494023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.218353 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a0706992809 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.940971017 +0000 UTC m=+3.309597370,LastTimestamp:2026-03-12 18:02:42.940971017 +0000 UTC m=+3.309597370,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.222206 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a0706aa32c2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.942087874 +0000 UTC m=+3.310714197,LastTimestamp:2026-03-12 18:02:42.942087874 +0000 UTC m=+3.310714197,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.227015 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a0706b30627 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:42.942666279 +0000 UTC m=+3.311292632,LastTimestamp:2026-03-12 18:02:42.942666279 +0000 UTC m=+3.311292632,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.231237 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a07161f395e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.201415518 +0000 UTC m=+3.570041871,LastTimestamp:2026-03-12 18:02:43.201415518 +0000 UTC m=+3.570041871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.235997 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a07163fde1f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.203554847 +0000 UTC m=+3.572181190,LastTimestamp:2026-03-12 18:02:43.203554847 +0000 UTC m=+3.572181190,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.240606 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c2a0716c78a36 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.212446262 +0000 UTC m=+3.581072595,LastTimestamp:2026-03-12 18:02:43.212446262 +0000 UTC m=+3.581072595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.244432 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a07172c68e5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.219056869 +0000 UTC m=+3.587683202,LastTimestamp:2026-03-12 18:02:43.219056869 +0000 UTC m=+3.587683202,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.251785 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a07173d3762 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.220158306 +0000 UTC m=+3.588784639,LastTimestamp:2026-03-12 18:02:43.220158306 +0000 UTC m=+3.588784639,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.256818 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a07221403f2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.402007538 +0000 UTC m=+3.770633871,LastTimestamp:2026-03-12 18:02:43.402007538 +0000 UTC m=+3.770633871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.260649 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a0722c5eb63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.413666659 +0000 UTC m=+3.782293012,LastTimestamp:2026-03-12 18:02:43.413666659 +0000 UTC m=+3.782293012,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.265933 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a0722df5fb7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.415334839 +0000 UTC m=+3.783961192,LastTimestamp:2026-03-12 18:02:43.415334839 +0000 UTC m=+3.783961192,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.271272 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0729b6b165 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.530109285 +0000 UTC m=+3.898735618,LastTimestamp:2026-03-12 18:02:43.530109285 +0000 UTC m=+3.898735618,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.276309 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a072dc99b9e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.598457758 +0000 UTC m=+3.967084101,LastTimestamp:2026-03-12 18:02:43.598457758 +0000 UTC m=+3.967084101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.280620 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a072e7bd965 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.610138981 +0000 UTC m=+3.978765324,LastTimestamp:2026-03-12 18:02:43.610138981 +0000 UTC m=+3.978765324,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.284970 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a07354dde28 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.724566056 +0000 UTC m=+4.093192389,LastTimestamp:2026-03-12 18:02:43.724566056 +0000 UTC m=+4.093192389,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.289931 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0735fba04a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.735953482 +0000 UTC m=+4.104579815,LastTimestamp:2026-03-12 18:02:43.735953482 +0000 UTC m=+4.104579815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.294614 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0766053795 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:44.541888405 +0000 UTC m=+4.910514738,LastTimestamp:2026-03-12 18:02:44.541888405 +0000 UTC m=+4.910514738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.299327 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0775962516 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:44.80304463 +0000 UTC m=+5.171670963,LastTimestamp:2026-03-12 18:02:44.80304463 +0000 UTC m=+5.171670963,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.303907 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a077666ba34 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:44.816714292 +0000 UTC m=+5.185340625,LastTimestamp:2026-03-12 18:02:44.816714292 +0000 UTC m=+5.185340625,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.308762 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0776763400 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:44.817728512 +0000 UTC m=+5.186354835,LastTimestamp:2026-03-12 18:02:44.817728512 +0000 UTC m=+5.186354835,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.313534 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0783c56a8d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.041023629 +0000 UTC m=+5.409649962,LastTimestamp:2026-03-12 18:02:45.041023629 +0000 UTC m=+5.409649962,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.316014 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0784e860c0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.060092096 +0000 UTC m=+5.428718429,LastTimestamp:2026-03-12 18:02:45.060092096 +0000 UTC m=+5.428718429,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.319583 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0784fd7f71 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.061476209 +0000 UTC m=+5.430102542,LastTimestamp:2026-03-12 18:02:45.061476209 +0000 UTC m=+5.430102542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.323790 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a0792a76c66 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.290716262 +0000 UTC m=+5.659342595,LastTimestamp:2026-03-12 18:02:45.290716262 +0000 UTC m=+5.659342595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.328228 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a079330c71a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.299717914 +0000 UTC m=+5.668344247,LastTimestamp:2026-03-12 18:02:45.299717914 +0000 UTC m=+5.668344247,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.332930 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a07933ec74e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.30063547 +0000 UTC m=+5.669261803,LastTimestamp:2026-03-12 18:02:45.30063547 +0000 UTC m=+5.669261803,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.336637 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a079e6cf416 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.488210966 +0000 UTC m=+5.856837309,LastTimestamp:2026-03-12 18:02:45.488210966 +0000 UTC m=+5.856837309,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.340683 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a079efe4905 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.497735429 +0000 UTC m=+5.866361752,LastTimestamp:2026-03-12 18:02:45.497735429 +0000 UTC m=+5.866361752,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.344671 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a079f0b4642 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.49858669 +0000 UTC m=+5.867213023,LastTimestamp:2026-03-12 18:02:45.49858669 +0000 UTC m=+5.867213023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.349144 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a07a96a1379 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.672571769 +0000 UTC m=+6.041198092,LastTimestamp:2026-03-12 18:02:45.672571769 +0000 UTC m=+6.041198092,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.353328 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c2a07aa444dce openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:45.68687355 +0000 UTC m=+6.055499883,LastTimestamp:2026-03-12 18:02:45.68687355 +0000 UTC m=+6.055499883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.359186 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 18:03:45 crc kubenswrapper[4926]: &Event{ObjectMeta:{kube-controller-manager-crc.189c2a09aa6e4922 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 18:03:45 crc kubenswrapper[4926]: body: Mar 12 18:03:45 crc kubenswrapper[4926]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.279559458 +0000 UTC m=+14.648185801,LastTimestamp:2026-03-12 18:02:54.279559458 +0000 UTC m=+14.648185801,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 18:03:45 crc kubenswrapper[4926]: > Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.362920 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a09aa6f404a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.27962273 +0000 UTC m=+14.648249073,LastTimestamp:2026-03-12 18:02:54.27962273 +0000 UTC m=+14.648249073,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.367260 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c2a0722df5fb7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a0722df5fb7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.415334839 +0000 UTC m=+3.783961192,LastTimestamp:2026-03-12 18:02:54.585949675 +0000 UTC m=+14.954576048,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.371419 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c2a072dc99b9e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a072dc99b9e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.598457758 +0000 UTC m=+3.967084101,LastTimestamp:2026-03-12 18:02:54.842625546 +0000 UTC m=+15.211251879,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.375663 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c2a072e7bd965\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a072e7bd965 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:43.610138981 +0000 UTC m=+3.978765324,LastTimestamp:2026-03-12 18:02:54.855078029 +0000 UTC m=+15.223704382,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.380564 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 12 18:03:45 crc kubenswrapper[4926]: &Event{ObjectMeta:{kube-apiserver-crc.189c2a09d19b0151 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 12 18:03:45 crc kubenswrapper[4926]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 18:03:45 crc kubenswrapper[4926]: Mar 12 18:03:45 crc kubenswrapper[4926]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.936801617 +0000 UTC m=+15.305427960,LastTimestamp:2026-03-12 18:02:54.936801617 +0000 UTC m=+15.305427960,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 18:03:45 crc kubenswrapper[4926]: > Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.384332 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a09d19c026f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.936867439 +0000 UTC m=+15.305493772,LastTimestamp:2026-03-12 18:02:54.936867439 +0000 UTC m=+15.305493772,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.389040 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c2a09d19b0151\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 12 18:03:45 crc kubenswrapper[4926]: &Event{ObjectMeta:{kube-apiserver-crc.189c2a09d19b0151 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 12 18:03:45 crc kubenswrapper[4926]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 18:03:45 crc kubenswrapper[4926]: Mar 12 18:03:45 crc kubenswrapper[4926]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.936801617 +0000 UTC m=+15.305427960,LastTimestamp:2026-03-12 18:02:54.943599313 +0000 UTC m=+15.312225656,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 18:03:45 crc kubenswrapper[4926]: > Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.393957 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c2a09d19c026f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c2a09d19c026f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.936867439 +0000 UTC m=+15.305493772,LastTimestamp:2026-03-12 18:02:54.943653735 +0000 UTC m=+15.312280068,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.400910 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a09aa6e4922\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 18:03:45 crc kubenswrapper[4926]: &Event{ObjectMeta:{kube-controller-manager-crc.189c2a09aa6e4922 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 18:03:45 crc kubenswrapper[4926]: body: Mar 12 18:03:45 crc kubenswrapper[4926]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.279559458 +0000 UTC m=+14.648185801,LastTimestamp:2026-03-12 18:03:04.279514677 +0000 UTC m=+24.648141010,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 18:03:45 crc kubenswrapper[4926]: > Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.405596 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a09aa6f404a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a09aa6f404a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.27962273 +0000 UTC m=+14.648249073,LastTimestamp:2026-03-12 18:03:04.279782097 +0000 UTC m=+24.648408430,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.409509 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 18:03:45 crc kubenswrapper[4926]: &Event{ObjectMeta:{kube-controller-manager-crc.189c2a0e1571ef35 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:46140->192.168.126.11:10357: read: connection reset by peer Mar 12 18:03:45 crc kubenswrapper[4926]: body: Mar 12 18:03:45 crc kubenswrapper[4926]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:03:13.254829877 +0000 UTC m=+33.623456220,LastTimestamp:2026-03-12 18:03:13.254829877 +0000 UTC m=+33.623456220,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 18:03:45 crc kubenswrapper[4926]: > Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.412903 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a0e1572babb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:46140->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:03:13.254881979 +0000 UTC m=+33.623508312,LastTimestamp:2026-03-12 18:03:13.254881979 +0000 UTC m=+33.623508312,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.418317 4926 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a0e15954f65 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:03:13.257148261 +0000 UTC m=+33.625774594,LastTimestamp:2026-03-12 18:03:13.257148261 +0000 UTC m=+33.625774594,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.420542 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a06b4e37904\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06b4e37904 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.5701097 +0000 UTC m=+1.938736043,LastTimestamp:2026-03-12 18:03:13.775076741 +0000 UTC m=+34.143703074,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: I0312 18:03:45.422406 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.422561 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a06c74187ff\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06c74187ff openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.878263807 +0000 UTC m=+2.246890180,LastTimestamp:2026-03-12 18:03:13.982368969 +0000 UTC m=+34.350995342,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.424164 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a06c7ece494\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a06c7ece494 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:41.889494164 +0000 UTC m=+2.258120527,LastTimestamp:2026-03-12 18:03:13.994832701 +0000 UTC m=+34.363459084,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.426943 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a09aa6e4922\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 18:03:45 crc kubenswrapper[4926]: &Event{ObjectMeta:{kube-controller-manager-crc.189c2a09aa6e4922 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 18:03:45 crc kubenswrapper[4926]: body: Mar 12 18:03:45 crc kubenswrapper[4926]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.279559458 +0000 UTC m=+14.648185801,LastTimestamp:2026-03-12 18:03:24.280145796 +0000 UTC m=+44.648772219,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 18:03:45 crc kubenswrapper[4926]: > Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.430721 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a09aa6f404a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c2a09aa6f404a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.27962273 +0000 UTC m=+14.648249073,LastTimestamp:2026-03-12 18:03:24.280240098 +0000 UTC m=+44.648866471,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:03:45 crc kubenswrapper[4926]: E0312 18:03:45.435362 4926 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c2a09aa6e4922\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 18:03:45 crc kubenswrapper[4926]: &Event{ObjectMeta:{kube-controller-manager-crc.189c2a09aa6e4922 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 18:03:45 crc kubenswrapper[4926]: body: Mar 12 18:03:45 crc kubenswrapper[4926]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:02:54.279559458 +0000 UTC m=+14.648185801,LastTimestamp:2026-03-12 18:03:34.279890169 +0000 UTC m=+54.648516512,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 18:03:45 crc kubenswrapper[4926]: > Mar 12 18:03:45 crc kubenswrapper[4926]: I0312 18:03:45.758110 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 12 18:03:45 crc kubenswrapper[4926]: I0312 18:03:45.760110 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:45 crc kubenswrapper[4926]: I0312 18:03:45.761205 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:45 crc kubenswrapper[4926]: I0312 18:03:45.761290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:45 crc kubenswrapper[4926]: I0312 18:03:45.761317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:46 crc kubenswrapper[4926]: I0312 18:03:46.427311 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.429016 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.489650 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.490985 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.491020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.491030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.491672 4926 scope.go:117] "RemoveContainer" containerID="237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.573218 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.573423 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.574754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.574790 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.574801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.766807 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.768330 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab"} Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.768493 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.769429 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.769476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:47 crc kubenswrapper[4926]: I0312 18:03:47.769487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:48 crc kubenswrapper[4926]: I0312 18:03:48.426919 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.427295 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.777504 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.778761 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.780867 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" exitCode=255 Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.780914 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab"} Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.780957 4926 scope.go:117] "RemoveContainer" containerID="237a0c1a00f49cb1e15478c5454fa052180c11a579cce5309cb7213bd78dff97" Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.781063 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.781816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.781839 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.781848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:49 crc kubenswrapper[4926]: I0312 18:03:49.782305 4926 scope.go:117] "RemoveContainer" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" Mar 12 18:03:49 crc kubenswrapper[4926]: E0312 18:03:49.782487 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:03:50 crc kubenswrapper[4926]: I0312 18:03:50.370007 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:50 crc kubenswrapper[4926]: I0312 18:03:50.371982 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:50 crc kubenswrapper[4926]: I0312 18:03:50.372259 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:50 crc kubenswrapper[4926]: I0312 18:03:50.372351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:50 crc kubenswrapper[4926]: I0312 18:03:50.372478 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 18:03:50 crc kubenswrapper[4926]: E0312 18:03:50.377894 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 18:03:50 crc kubenswrapper[4926]: E0312 18:03:50.378942 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 18:03:50 crc kubenswrapper[4926]: I0312 18:03:50.427379 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:50 crc kubenswrapper[4926]: E0312 18:03:50.565263 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 18:03:50 crc kubenswrapper[4926]: I0312 18:03:50.784885 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.229156 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.229407 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.230585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.230617 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.230627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.231081 4926 scope.go:117] "RemoveContainer" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" Mar 12 18:03:51 crc kubenswrapper[4926]: E0312 18:03:51.231244 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.279852 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.280070 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.281088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.281115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.281128 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.329244 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.337296 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.426508 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.712682 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.741361 4926 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.788858 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.788901 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.790095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.790118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.790126 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.790188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.790204 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.790214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:51 crc kubenswrapper[4926]: I0312 18:03:51.790741 4926 scope.go:117] "RemoveContainer" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" Mar 12 18:03:51 crc kubenswrapper[4926]: E0312 18:03:51.790909 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:03:52 crc kubenswrapper[4926]: I0312 18:03:52.427316 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:53 crc kubenswrapper[4926]: I0312 18:03:53.426735 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:54 crc kubenswrapper[4926]: I0312 18:03:54.428152 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:55 crc kubenswrapper[4926]: I0312 18:03:55.429840 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:55 crc kubenswrapper[4926]: W0312 18:03:55.711958 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 12 18:03:55 crc kubenswrapper[4926]: E0312 18:03:55.712029 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 18:03:55 crc kubenswrapper[4926]: W0312 18:03:55.880685 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 12 18:03:55 crc kubenswrapper[4926]: E0312 18:03:55.880741 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 18:03:56 crc kubenswrapper[4926]: I0312 18:03:56.025652 4926 csr.go:261] certificate signing request csr-f7zxj is approved, waiting to be issued Mar 12 18:03:56 crc kubenswrapper[4926]: I0312 18:03:56.037176 4926 csr.go:257] certificate signing request csr-f7zxj is issued Mar 12 18:03:56 crc kubenswrapper[4926]: I0312 18:03:56.051921 4926 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 12 18:03:56 crc kubenswrapper[4926]: I0312 18:03:56.250761 4926 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.038531 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-19 20:51:38.749459399 +0000 UTC Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.038590 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7514h47m41.710878482s for next certificate rotation Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.379648 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.380842 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.380891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.380901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.380989 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.389591 4926 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.389914 4926 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.389949 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.396589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.396627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.396636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.396650 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.396659 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:03:57Z","lastTransitionTime":"2026-03-12T18:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.414907 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.425912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.425955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.425965 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.425979 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.425990 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:03:57Z","lastTransitionTime":"2026-03-12T18:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.436768 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.445901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.445956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.445970 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.445990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.446004 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:03:57Z","lastTransitionTime":"2026-03-12T18:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.460406 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.468299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.468346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.468356 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.468372 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.468383 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:03:57Z","lastTransitionTime":"2026-03-12T18:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.483679 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.484044 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.484104 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.576308 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.576532 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.577840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.577908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:03:57 crc kubenswrapper[4926]: I0312 18:03:57.577926 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.584427 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.684810 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.785359 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.886527 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:57 crc kubenswrapper[4926]: E0312 18:03:57.987644 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.088073 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.188777 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.289497 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.390002 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.490787 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.591885 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.692628 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.793598 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.894270 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:58 crc kubenswrapper[4926]: E0312 18:03:58.995150 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.095722 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.195863 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.296380 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.397559 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.497702 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.599255 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.699498 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.800246 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:03:59 crc kubenswrapper[4926]: E0312 18:03:59.901208 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.001830 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.102937 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.203931 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.304594 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.405049 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.505720 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.565732 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.606153 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.706977 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.807866 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:00 crc kubenswrapper[4926]: E0312 18:04:00.908604 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.008911 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.110002 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.211095 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.311628 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.411755 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.512761 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.613029 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.713933 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.814356 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:01 crc kubenswrapper[4926]: E0312 18:04:01.915413 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.016349 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.117302 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.218225 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.318635 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.418814 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.519228 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.619341 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.720319 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.821065 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:02 crc kubenswrapper[4926]: E0312 18:04:02.921867 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.022309 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.123401 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.223927 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.324694 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.425129 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.525326 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.625764 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.725953 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.827154 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:03 crc kubenswrapper[4926]: E0312 18:04:03.928304 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.029525 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.129806 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.230389 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.331533 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.431687 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: I0312 18:04:04.489612 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:04:04 crc kubenswrapper[4926]: I0312 18:04:04.490828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:04 crc kubenswrapper[4926]: I0312 18:04:04.490860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:04 crc kubenswrapper[4926]: I0312 18:04:04.490868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:04 crc kubenswrapper[4926]: I0312 18:04:04.491373 4926 scope.go:117] "RemoveContainer" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.491552 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.532006 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.632120 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.732262 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.832386 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:04 crc kubenswrapper[4926]: E0312 18:04:04.933713 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.034493 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.135542 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.236785 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.337883 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.438045 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.538967 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.639432 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.740478 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.841523 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:05 crc kubenswrapper[4926]: E0312 18:04:05.942680 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.043362 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.143683 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.244712 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.344946 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.445661 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.546335 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.646543 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.747104 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.848211 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:06 crc kubenswrapper[4926]: E0312 18:04:06.949245 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.050144 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.151173 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.251347 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.351844 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.452518 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.552638 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.653329 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.753795 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.854195 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.873573 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.879875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.879988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.880017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.880057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.880086 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:07Z","lastTransitionTime":"2026-03-12T18:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.898876 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.910614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.910865 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.910941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.911036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.911107 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:07Z","lastTransitionTime":"2026-03-12T18:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.922196 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.933768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.933935 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.934040 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.934086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.934123 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:07Z","lastTransitionTime":"2026-03-12T18:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.951276 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.962949 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.962994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.963005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.963030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:07 crc kubenswrapper[4926]: I0312 18:04:07.963041 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:07Z","lastTransitionTime":"2026-03-12T18:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.975038 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.975179 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:04:07 crc kubenswrapper[4926]: E0312 18:04:07.975207 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.075725 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.176643 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.276754 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.377855 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.478213 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.578748 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.678848 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.779597 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.880014 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:08 crc kubenswrapper[4926]: E0312 18:04:08.980770 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.081488 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.182339 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.283512 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.384587 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.485777 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.586992 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.687902 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.789103 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.889697 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:09 crc kubenswrapper[4926]: E0312 18:04:09.990805 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.091969 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.192171 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.292818 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.393521 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: I0312 18:04:10.489386 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:04:10 crc kubenswrapper[4926]: I0312 18:04:10.491244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:10 crc kubenswrapper[4926]: I0312 18:04:10.491318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:10 crc kubenswrapper[4926]: I0312 18:04:10.491341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.494376 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.566160 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.594551 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.695057 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.795563 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.896535 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:10 crc kubenswrapper[4926]: E0312 18:04:10.997692 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.098776 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.199947 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.300260 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.400950 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.502045 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.603013 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.704224 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.805607 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:11 crc kubenswrapper[4926]: E0312 18:04:11.906759 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.007519 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.108104 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.209017 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.309396 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.410338 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.510456 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.610768 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.711869 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.812619 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:12 crc kubenswrapper[4926]: E0312 18:04:12.913726 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.014846 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.115929 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.217124 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.318211 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.418683 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.519500 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.619640 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.720699 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.821052 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:13 crc kubenswrapper[4926]: E0312 18:04:13.921665 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.022350 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.123306 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.224229 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.324595 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.424723 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.524825 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.625369 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.725729 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.826194 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:14 crc kubenswrapper[4926]: E0312 18:04:14.926904 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.028026 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.128504 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.229085 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.329519 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.430168 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.530360 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.631038 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.731998 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.832345 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:15 crc kubenswrapper[4926]: E0312 18:04:15.933332 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.034311 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.135505 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.235588 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.336087 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.436592 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: I0312 18:04:16.489881 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:04:16 crc kubenswrapper[4926]: I0312 18:04:16.491024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:16 crc kubenswrapper[4926]: I0312 18:04:16.491083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:16 crc kubenswrapper[4926]: I0312 18:04:16.491096 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:16 crc kubenswrapper[4926]: I0312 18:04:16.491759 4926 scope.go:117] "RemoveContainer" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.491926 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 18:04:16 crc kubenswrapper[4926]: I0312 18:04:16.514585 4926 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.537341 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.637760 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.738222 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.839378 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:16 crc kubenswrapper[4926]: E0312 18:04:16.939946 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.040820 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.141979 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.242307 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.342876 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.443682 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.544489 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.644589 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.744895 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.845605 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:17 crc kubenswrapper[4926]: E0312 18:04:17.946759 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.047907 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.148383 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.248527 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.349018 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.350152 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.354510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.354575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.354584 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.354600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.354609 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:18Z","lastTransitionTime":"2026-03-12T18:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.371074 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.375161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.375211 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.375222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.375240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.375255 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:18Z","lastTransitionTime":"2026-03-12T18:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.389503 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.394099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.394145 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.394179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.394197 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.394207 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:18Z","lastTransitionTime":"2026-03-12T18:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.404101 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.408481 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.408543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.408555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.408570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:18 crc kubenswrapper[4926]: I0312 18:04:18.408579 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:18Z","lastTransitionTime":"2026-03-12T18:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.418834 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.418945 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.449150 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.549609 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.650108 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.750254 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.850839 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:18 crc kubenswrapper[4926]: E0312 18:04:18.951320 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.052297 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.152501 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.253403 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.353751 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.454586 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.555238 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.656413 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.757402 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.857821 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:19 crc kubenswrapper[4926]: E0312 18:04:19.958756 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.059120 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.160019 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.260324 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.361000 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.461155 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.562136 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.566304 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.663240 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.763583 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.864802 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:20 crc kubenswrapper[4926]: E0312 18:04:20.965237 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.065555 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.166570 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.267648 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.368748 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.469058 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.570035 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.670392 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.771232 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.871708 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:21 crc kubenswrapper[4926]: E0312 18:04:21.972241 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.072427 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.173093 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.273506 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.373970 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.474390 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.574749 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.675856 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.776895 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.877773 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:22 crc kubenswrapper[4926]: E0312 18:04:22.978091 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.078307 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.179303 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.280275 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.380940 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.482113 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.582635 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.683762 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.784044 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.884667 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:23 crc kubenswrapper[4926]: E0312 18:04:23.984826 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.085839 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.186111 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.287165 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.387664 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.488864 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.589542 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.690104 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.790596 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.891363 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:24 crc kubenswrapper[4926]: E0312 18:04:24.991964 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.092691 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.193650 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.294759 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.395582 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: I0312 18:04:25.489904 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 18:04:25 crc kubenswrapper[4926]: I0312 18:04:25.493990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:25 crc kubenswrapper[4926]: I0312 18:04:25.494033 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:25 crc kubenswrapper[4926]: I0312 18:04:25.494046 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.496349 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.596676 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.697115 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.797808 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.898702 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:25 crc kubenswrapper[4926]: E0312 18:04:25.999801 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.100653 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.177151 4926 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.203941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.203982 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.203992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.204008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.204020 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.306712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.306776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.306789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.306805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.306818 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.408887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.408930 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.408941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.408959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.408971 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.440880 4926 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.466162 4926 apiserver.go:52] "Watching apiserver" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.470733 4926 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.471116 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-machine-config-operator/machine-config-daemon-hmdg8","openshift-multus/multus-xwqvl","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-dns/node-resolver-f9vxh","openshift-multus/multus-additional-cni-plugins-srh42","openshift-ovn-kubernetes/ovnkube-node-zlfmg"] Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.471547 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.471555 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.471605 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.471667 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.471679 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.471721 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.471738 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.472386 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.472571 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.472484 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.472954 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.473322 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.473936 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.474102 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.476888 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.476891 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.477126 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.477230 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.477321 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.477376 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.477516 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.477956 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.477984 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478117 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478138 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478218 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478241 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478316 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478336 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478428 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478578 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478666 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478684 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.478722 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479054 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479088 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479058 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479139 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479166 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479199 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479267 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479272 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479366 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479386 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.479488 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.501121 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.512026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.512076 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.512087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.512104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.512115 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.514109 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.524140 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.530719 4926 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.537619 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.549003 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.557892 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.567129 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.572372 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.572538 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.572667 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.572805 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.572957 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573091 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573198 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573291 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573383 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573515 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.572714 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573763 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573703 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573814 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573826 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.572989 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573850 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573872 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573225 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573896 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573922 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573944 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573976 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574001 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574018 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574035 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574056 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574079 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574098 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573182 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573400 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573617 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573626 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.573654 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574090 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574118 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574245 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574291 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574242 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574342 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574411 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574448 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574456 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574480 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574506 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574531 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574554 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574575 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574596 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574619 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574728 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574780 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574804 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574826 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574849 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574875 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574898 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574921 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574943 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574965 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574989 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575011 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575036 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575058 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575080 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575103 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575126 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575147 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575169 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575190 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575211 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575233 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575255 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575277 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575299 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575321 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575345 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575366 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575388 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575414 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575439 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575502 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575525 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575546 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575567 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575589 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575640 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575661 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575683 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575705 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575728 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575750 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575772 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575796 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575818 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575843 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575869 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575894 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575916 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575938 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575960 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575981 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576004 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576028 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576051 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576074 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576096 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576118 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576141 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576161 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576184 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576222 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576245 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576267 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576297 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576337 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576379 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576408 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576446 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576502 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576538 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576567 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576589 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576613 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576638 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576684 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576712 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576745 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576774 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576806 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576833 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576861 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576901 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576950 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576996 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577048 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577101 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577136 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577169 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577202 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577235 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577268 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577301 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577337 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577374 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577419 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577540 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577593 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577644 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577678 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577718 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577789 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577827 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577878 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577928 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577966 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578001 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578034 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578072 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578108 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578140 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578171 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578205 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578555 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578604 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578642 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578695 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578770 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578759 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578824 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578860 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578894 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578927 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578960 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578995 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579056 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579122 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579157 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579189 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579223 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579258 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579292 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579330 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579368 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579403 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579443 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579531 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579570 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579605 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579640 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579675 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579718 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579787 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579823 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579858 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579892 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579926 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579961 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579997 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580033 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580068 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580103 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580137 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580172 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580211 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580247 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580285 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580324 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580359 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580438 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5a53ef4-c701-457f-9cf2-85819bf04d1a-cni-binary-copy\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580505 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-etc-kubernetes\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580541 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-slash\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580574 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-var-lib-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580609 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-env-overrides\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580643 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-cnibin\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580675 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-k8s-cni-cncf-io\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580708 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580743 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbt59\" (UniqueName: \"kubernetes.io/projected/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-kube-api-access-nbt59\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580775 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-netns\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580807 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-kubelet\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580840 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-systemd-units\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580873 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8t62\" (UniqueName: \"kubernetes.io/projected/f7b34559-da2f-4796-8f3f-c56b2725c464-kube-api-access-v8t62\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-socket-dir-parent\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580944 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-multus-certs\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580977 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7bdm\" (UniqueName: \"kubernetes.io/projected/594c806d-dd79-41ce-8e3a-a33d42bf0f7e-kube-api-access-x7bdm\") pod \"node-resolver-f9vxh\" (UID: \"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\") " pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581010 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cnibin\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581078 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581114 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-config\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581148 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f7b34559-da2f-4796-8f3f-c56b2725c464-rootfs\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581178 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-hostroot\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581210 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581242 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-script-lib\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581274 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-netns\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581305 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-netd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581362 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581401 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581447 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581500 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-conf-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-systemd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581569 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-ovn-kubernetes\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581610 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581647 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581681 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-os-release\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581807 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt6hw\" (UniqueName: \"kubernetes.io/projected/d5a53ef4-c701-457f-9cf2-85819bf04d1a-kube-api-access-bt6hw\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581977 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-system-cni-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582018 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-cni-bin\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582057 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cni-binary-copy\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582092 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-node-log\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582123 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovn-node-metrics-cert\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582163 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582200 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-cni-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582234 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-log-socket\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582273 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582312 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582449 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582509 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582551 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t5dd\" (UniqueName: \"kubernetes.io/projected/bc33af41-5aa0-4254-ac75-69433d5f4ce9-kube-api-access-4t5dd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582589 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-kubelet\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582622 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-daemon-config\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582655 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-system-cni-dir\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582686 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-bin\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582722 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f7b34559-da2f-4796-8f3f-c56b2725c464-proxy-tls\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582889 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/594c806d-dd79-41ce-8e3a-a33d42bf0f7e-hosts-file\") pod \"node-resolver-f9vxh\" (UID: \"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\") " pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582958 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-ovn\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582999 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583035 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583070 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583106 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-cni-multus\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583150 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-etc-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583185 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f7b34559-da2f-4796-8f3f-c56b2725c464-mcd-auth-proxy-config\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583225 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-os-release\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583306 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583331 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583355 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583377 4926 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583399 4926 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574480 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583526 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583548 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583696 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583705 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583756 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583804 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.584166 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.584449 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.584726 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.585054 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.585081 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.585104 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.586483 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.586586 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.586650 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.586985 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583422 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587073 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587105 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587131 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587157 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587181 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587208 4926 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587233 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587259 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574630 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574686 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574720 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574786 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587535 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.587714 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:27.087669238 +0000 UTC m=+107.456295631 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587765 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587956 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.587835 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.588172 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.588840 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.588866 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.588823 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.589536 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:04:27.089522237 +0000 UTC m=+107.458148570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.589683 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.574849 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575063 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575233 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.575189 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576390 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576717 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576711 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576734 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.590111 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.576905 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577197 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577228 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577315 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577256 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577076 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577683 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.577994 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578000 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578030 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578231 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578521 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578542 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578647 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578628 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579203 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.578920 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579767 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579835 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.579894 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.580761 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581017 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581052 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581359 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581372 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581415 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581435 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581641 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581713 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581735 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581870 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.581871 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582200 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582287 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582327 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582744 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.582880 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583062 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583074 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583290 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583333 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.583392 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.589992 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.590140 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.590432 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.590445 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.590512 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591022 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591095 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591135 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591221 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.591335 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:27.091254171 +0000 UTC m=+107.459880564 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.590735 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591612 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591650 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591670 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.591932 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.592227 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.592269 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.592552 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.592560 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.592877 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.592957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.593064 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.593566 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.593572 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.593630 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.593723 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.593842 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594050 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594066 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594170 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594230 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594342 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594388 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594431 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.594621 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.595371 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.595737 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.595727 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.595940 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.596122 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.596126 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.596337 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.596503 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.596517 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.596937 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.597334 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.598177 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.598508 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.599077 4926 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.599730 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.600262 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.600279 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.602496 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.603199 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.603541 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.603791 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.604022 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.604729 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.604875 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.605291 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.605287 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.605856 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.605936 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.605950 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.605980 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.605997 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.606138 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:27.106119479 +0000 UTC m=+107.474745842 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.606398 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.607004 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.607389 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.607715 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.609969 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.611257 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.611344 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.611427 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.611641 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.611712 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.611708 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.611843 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.612131 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.612203 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.612394 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614157 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614736 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614856 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.614861 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.615613 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.616983 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.617012 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.617027 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:26 crc kubenswrapper[4926]: E0312 18:04:26.617085 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:27.117061894 +0000 UTC m=+107.485688307 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.617499 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.618747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.618771 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.618921 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.618971 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.619086 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.619108 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.619143 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.619666 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.621112 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.621574 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.621801 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.621933 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.623184 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.623537 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.623419 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.624178 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.624235 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.624330 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.624455 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.627761 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.628397 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.628662 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.629611 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.629867 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.642027 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.645215 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.654647 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.655214 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.669054 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.671966 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.673564 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.677764 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t5dd\" (UniqueName: \"kubernetes.io/projected/bc33af41-5aa0-4254-ac75-69433d5f4ce9-kube-api-access-4t5dd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688378 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-kubelet\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688400 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-daemon-config\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688422 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-system-cni-dir\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688445 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-bin\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688483 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f7b34559-da2f-4796-8f3f-c56b2725c464-proxy-tls\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688498 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-cni-multus\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688513 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/594c806d-dd79-41ce-8e3a-a33d42bf0f7e-hosts-file\") pod \"node-resolver-f9vxh\" (UID: \"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\") " pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688528 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-ovn\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688554 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-os-release\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688566 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-etc-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688580 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f7b34559-da2f-4796-8f3f-c56b2725c464-mcd-auth-proxy-config\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688594 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5a53ef4-c701-457f-9cf2-85819bf04d1a-cni-binary-copy\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688612 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-etc-kubernetes\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688879 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-slash\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688926 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-var-lib-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688959 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-env-overrides\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688988 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-kubelet\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689057 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-ovn\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689057 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-cnibin\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689075 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-cni-multus\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689102 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-k8s-cni-cncf-io\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689105 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-cnibin\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689153 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-k8s-cni-cncf-io\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/594c806d-dd79-41ce-8e3a-a33d42bf0f7e-hosts-file\") pod \"node-resolver-f9vxh\" (UID: \"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\") " pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689283 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-etc-kubernetes\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689335 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-slash\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689354 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-os-release\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689384 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-bin\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.688624 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-system-cni-dir\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-etc-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689489 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-var-lib-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689634 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689757 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbt59\" (UniqueName: \"kubernetes.io/projected/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-kube-api-access-nbt59\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689804 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-netns\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689888 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-kubelet\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689928 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-systemd-units\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.689969 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8t62\" (UniqueName: \"kubernetes.io/projected/f7b34559-da2f-4796-8f3f-c56b2725c464-kube-api-access-v8t62\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690065 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-config\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-socket-dir-parent\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690114 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-daemon-config\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690248 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-multus-certs\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690290 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7bdm\" (UniqueName: \"kubernetes.io/projected/594c806d-dd79-41ce-8e3a-a33d42bf0f7e-kube-api-access-x7bdm\") pod \"node-resolver-f9vxh\" (UID: \"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\") " pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690353 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cnibin\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690493 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-env-overrides\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690582 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-systemd-units\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690590 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-netns\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690661 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690688 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690697 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-multus-certs\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690771 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f7b34559-da2f-4796-8f3f-c56b2725c464-rootfs\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690834 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-hostroot\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.690937 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691041 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-script-lib\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691132 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-netns\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691215 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-netd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-run-netns\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691401 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-netd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691476 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f7b34559-da2f-4796-8f3f-c56b2725c464-rootfs\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691546 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-openvswitch\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691616 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691707 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5a53ef4-c701-457f-9cf2-85819bf04d1a-cni-binary-copy\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-kubelet\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691843 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691877 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-hostroot\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691936 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cnibin\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691937 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-script-lib\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691966 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-socket-dir-parent\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692026 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-conf-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.691986 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-conf-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692092 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-os-release\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692163 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-systemd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692220 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-ovn-kubernetes\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-systemd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692316 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692317 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692358 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-os-release\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-ovn-kubernetes\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692428 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f7b34559-da2f-4796-8f3f-c56b2725c464-mcd-auth-proxy-config\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692490 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt6hw\" (UniqueName: \"kubernetes.io/projected/d5a53ef4-c701-457f-9cf2-85819bf04d1a-kube-api-access-bt6hw\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.692529 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-system-cni-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694180 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-config\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694250 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-cni-bin\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-host-var-lib-cni-bin\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694319 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-system-cni-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694363 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cni-binary-copy\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694546 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-node-log\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694630 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-node-log\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694696 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovn-node-metrics-cert\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.694766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-cni-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695664 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-log-socket\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695712 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-log-socket\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695672 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5a53ef4-c701-457f-9cf2-85819bf04d1a-multus-cni-dir\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695741 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695868 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695963 4926 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695983 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.695998 4926 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696017 4926 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696029 4926 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696041 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696062 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696075 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696088 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696087 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-cni-binary-copy\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696099 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696156 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696174 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696195 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696238 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696253 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696267 4926 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696285 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696299 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696311 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696324 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696341 4926 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696354 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696366 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696378 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696394 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696406 4926 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696417 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696462 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696475 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696488 4926 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696501 4926 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696517 4926 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696530 4926 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696585 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696634 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696656 4926 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696669 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696680 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696695 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696717 4926 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696728 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696797 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696835 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696847 4926 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696859 4926 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696888 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696900 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696911 4926 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696922 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696936 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696947 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696957 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696969 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696978 4926 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.696989 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697002 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697017 4926 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697029 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697039 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697049 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697061 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697071 4926 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697081 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697093 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697107 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697116 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697127 4926 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697144 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697155 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697167 4926 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697180 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697199 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697227 4926 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697239 4926 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697250 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697264 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697274 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697284 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697298 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697310 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697325 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697337 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697352 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697361 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697371 4926 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697380 4926 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697393 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697402 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697412 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697423 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697444 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697470 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697494 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697511 4926 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697522 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697533 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697546 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697561 4926 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697571 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697582 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697601 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697614 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697632 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697642 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697656 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697669 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697679 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697689 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697710 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697722 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697733 4926 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697744 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697765 4926 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697774 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697784 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697793 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697807 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697817 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697827 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697839 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697850 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697861 4926 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697875 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697887 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697899 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697912 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697922 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697936 4926 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697947 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697956 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697967 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697979 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697989 4926 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.697999 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698011 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698021 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698035 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698044 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698055 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698064 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698074 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698088 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698103 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698104 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f7b34559-da2f-4796-8f3f-c56b2725c464-proxy-tls\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698112 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698187 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698207 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698237 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698255 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698269 4926 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698282 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698299 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698310 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698323 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698339 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698351 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698363 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698375 4926 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698401 4926 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698414 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698425 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698449 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698489 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698503 4926 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698518 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698535 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698548 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698561 4926 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698573 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698588 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698601 4926 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698613 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698625 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698666 4926 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698681 4926 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698695 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698711 4926 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698779 4926 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698799 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698812 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698832 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.698957 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovn-node-metrics-cert\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.704687 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t5dd\" (UniqueName: \"kubernetes.io/projected/bc33af41-5aa0-4254-ac75-69433d5f4ce9-kube-api-access-4t5dd\") pod \"ovnkube-node-zlfmg\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.706223 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbt59\" (UniqueName: \"kubernetes.io/projected/1d37aa11-8fa5-4eb3-8edd-6f71523623b5-kube-api-access-nbt59\") pod \"multus-additional-cni-plugins-srh42\" (UID: \"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\") " pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.708197 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt6hw\" (UniqueName: \"kubernetes.io/projected/d5a53ef4-c701-457f-9cf2-85819bf04d1a-kube-api-access-bt6hw\") pod \"multus-xwqvl\" (UID: \"d5a53ef4-c701-457f-9cf2-85819bf04d1a\") " pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.712362 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8t62\" (UniqueName: \"kubernetes.io/projected/f7b34559-da2f-4796-8f3f-c56b2725c464-kube-api-access-v8t62\") pod \"machine-config-daemon-hmdg8\" (UID: \"f7b34559-da2f-4796-8f3f-c56b2725c464\") " pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.715631 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7bdm\" (UniqueName: \"kubernetes.io/projected/594c806d-dd79-41ce-8e3a-a33d42bf0f7e-kube-api-access-x7bdm\") pod \"node-resolver-f9vxh\" (UID: \"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\") " pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.716902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.716927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.716936 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.716950 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.716960 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.787372 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.795678 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.806349 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 18:04:26 crc kubenswrapper[4926]: W0312 18:04:26.813215 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-72f0695c2edf73654195d549a62c8fcccfaf8fc5295e2c261ac95fcbdc4e578b WatchSource:0}: Error finding container 72f0695c2edf73654195d549a62c8fcccfaf8fc5295e2c261ac95fcbdc4e578b: Status 404 returned error can't find the container with id 72f0695c2edf73654195d549a62c8fcccfaf8fc5295e2c261ac95fcbdc4e578b Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.816397 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.820482 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.820515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.820526 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.820543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.820557 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.824728 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xwqvl" Mar 12 18:04:26 crc kubenswrapper[4926]: W0312 18:04:26.826097 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-8d53cdc4e48c4d7f3859e37e80091956822e666ab97637824a53ad05c3b41920 WatchSource:0}: Error finding container 8d53cdc4e48c4d7f3859e37e80091956822e666ab97637824a53ad05c3b41920: Status 404 returned error can't find the container with id 8d53cdc4e48c4d7f3859e37e80091956822e666ab97637824a53ad05c3b41920 Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.832055 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.839705 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-f9vxh" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.846775 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-srh42" Mar 12 18:04:26 crc kubenswrapper[4926]: W0312 18:04:26.847267 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7b34559_da2f_4796_8f3f_c56b2725c464.slice/crio-f4b2e4a544758c1263364c0df7e07d77c8bc4370a470d837ea6efae07158abf9 WatchSource:0}: Error finding container f4b2e4a544758c1263364c0df7e07d77c8bc4370a470d837ea6efae07158abf9: Status 404 returned error can't find the container with id f4b2e4a544758c1263364c0df7e07d77c8bc4370a470d837ea6efae07158abf9 Mar 12 18:04:26 crc kubenswrapper[4926]: W0312 18:04:26.853638 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5a53ef4_c701_457f_9cf2_85819bf04d1a.slice/crio-ad259f0c6d06a3faa766bd42496c02cc7bc98b9a5ebc84ff9eeb4ed17814b7e7 WatchSource:0}: Error finding container ad259f0c6d06a3faa766bd42496c02cc7bc98b9a5ebc84ff9eeb4ed17814b7e7: Status 404 returned error can't find the container with id ad259f0c6d06a3faa766bd42496c02cc7bc98b9a5ebc84ff9eeb4ed17814b7e7 Mar 12 18:04:26 crc kubenswrapper[4926]: W0312 18:04:26.873820 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod594c806d_dd79_41ce_8e3a_a33d42bf0f7e.slice/crio-12064d7968642d5fcd14005252eca5815ef36389939f49bd05f0edaa57a27ba8 WatchSource:0}: Error finding container 12064d7968642d5fcd14005252eca5815ef36389939f49bd05f0edaa57a27ba8: Status 404 returned error can't find the container with id 12064d7968642d5fcd14005252eca5815ef36389939f49bd05f0edaa57a27ba8 Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.881896 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1f450f062448c5e8fdb1b02bd5b83f974900c9bf14e3078dff971588b7fe4341"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.883898 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-f9vxh" event={"ID":"594c806d-dd79-41ce-8e3a-a33d42bf0f7e","Type":"ContainerStarted","Data":"12064d7968642d5fcd14005252eca5815ef36389939f49bd05f0edaa57a27ba8"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.885291 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"f4b2e4a544758c1263364c0df7e07d77c8bc4370a470d837ea6efae07158abf9"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.886478 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"72f0695c2edf73654195d549a62c8fcccfaf8fc5295e2c261ac95fcbdc4e578b"} Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.890087 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerStarted","Data":"ad259f0c6d06a3faa766bd42496c02cc7bc98b9a5ebc84ff9eeb4ed17814b7e7"} Mar 12 18:04:26 crc kubenswrapper[4926]: W0312 18:04:26.891179 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc33af41_5aa0_4254_ac75_69433d5f4ce9.slice/crio-30c6b754be03f1fd88325819e9237821f35a5d4ca5f0ad8545574e86e65cadf9 WatchSource:0}: Error finding container 30c6b754be03f1fd88325819e9237821f35a5d4ca5f0ad8545574e86e65cadf9: Status 404 returned error can't find the container with id 30c6b754be03f1fd88325819e9237821f35a5d4ca5f0ad8545574e86e65cadf9 Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.891761 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8d53cdc4e48c4d7f3859e37e80091956822e666ab97637824a53ad05c3b41920"} Mar 12 18:04:26 crc kubenswrapper[4926]: W0312 18:04:26.895337 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d37aa11_8fa5_4eb3_8edd_6f71523623b5.slice/crio-6834c4d42bb80ef9b2cbc21221a1f2bd8626677d5b221112fb9686702721bfff WatchSource:0}: Error finding container 6834c4d42bb80ef9b2cbc21221a1f2bd8626677d5b221112fb9686702721bfff: Status 404 returned error can't find the container with id 6834c4d42bb80ef9b2cbc21221a1f2bd8626677d5b221112fb9686702721bfff Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.924022 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.924070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.924083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.924103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:26 crc kubenswrapper[4926]: I0312 18:04:26.924116 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:26Z","lastTransitionTime":"2026-03-12T18:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.026504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.026548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.026557 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.026570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.026579 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.102027 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.102154 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.102186 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.102272 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.102339 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:28.102313075 +0000 UTC m=+108.470939418 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.102476 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.102553 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:28.102534653 +0000 UTC m=+108.471160986 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.102618 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:04:28.102573804 +0000 UTC m=+108.471200137 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.128904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.128939 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.128950 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.128966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.128978 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.202900 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.202970 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203071 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203085 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203095 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203106 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203133 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203140 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:28.20312775 +0000 UTC m=+108.571754083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203147 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:27 crc kubenswrapper[4926]: E0312 18:04:27.203195 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:28.203179082 +0000 UTC m=+108.571805415 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.231715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.231761 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.231772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.231791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.231803 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.334860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.334902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.334917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.334933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.334944 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.437670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.437722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.437733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.437750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.437763 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.540861 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.540906 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.540918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.540935 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.540949 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.643263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.643298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.643306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.643319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.643327 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.745663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.745726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.745747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.745775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.745796 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.848770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.848834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.848854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.848880 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.848898 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.900702 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.900750 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.904042 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.904074 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.907259 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.910578 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerStarted","Data":"54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.912514 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d37aa11-8fa5-4eb3-8edd-6f71523623b5" containerID="b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818" exitCode=0 Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.912580 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerDied","Data":"b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.912601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerStarted","Data":"6834c4d42bb80ef9b2cbc21221a1f2bd8626677d5b221112fb9686702721bfff"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.915983 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe" exitCode=0 Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.916037 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.916053 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"30c6b754be03f1fd88325819e9237821f35a5d4ca5f0ad8545574e86e65cadf9"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.918533 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-f9vxh" event={"ID":"594c806d-dd79-41ce-8e3a-a33d42bf0f7e","Type":"ContainerStarted","Data":"690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.929995 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.952654 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.953672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.953707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.953721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.953740 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.953754 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:27Z","lastTransitionTime":"2026-03-12T18:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.968971 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:27 crc kubenswrapper[4926]: I0312 18:04:27.989582 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.004608 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.021831 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.043155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.055734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.055777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.055788 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.055805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.055818 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.066271 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.080090 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.096825 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.108880 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.111696 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.111886 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.111937 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.111993 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:04:30.111968621 +0000 UTC m=+110.480595104 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.111999 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.112044 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.112067 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:30.112057304 +0000 UTC m=+110.480683857 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.112111 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:30.112091645 +0000 UTC m=+110.480717978 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.120236 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.133408 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.146015 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.158241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.158276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.158287 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.158303 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.158314 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.160081 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.171280 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.181937 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.192290 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.211040 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.212389 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.212462 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212568 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212587 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212599 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212635 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:30.212622241 +0000 UTC m=+110.581248574 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212568 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212836 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212846 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.212872 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:30.212863488 +0000 UTC m=+110.581489821 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.229673 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.242115 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.253676 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.260248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.260286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.260296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.260313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.260322 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.362644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.362690 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.362706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.362728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.362744 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.469288 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.469687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.469699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.469717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.469731 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.489643 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.489757 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.489811 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.489882 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.495419 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.495699 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.499252 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.500307 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.501672 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.502424 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.503500 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.504134 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.504862 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.506424 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.513772 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.514606 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.516976 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.518040 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.519208 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.520052 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.520685 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.521849 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.522750 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.523822 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.524718 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.525726 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.526831 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.527547 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.529347 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.530193 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.530831 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.532051 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.533309 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.533941 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.534737 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.536059 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.536581 4926 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.536696 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.538747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.538781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.538789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.538803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.538813 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.538977 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.539635 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.540379 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.542019 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.542798 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.543850 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.544648 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.545945 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.546595 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.548297 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.549112 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.550364 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.550970 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.552062 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.552746 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.554071 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.554724 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.555896 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.556585 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.557256 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.557738 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.559140 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.559991 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.563823 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.563882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.563902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.563925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.563941 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.582331 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.586734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.586767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.586779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.586796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.586808 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.598320 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.602663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.602721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.602737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.602761 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.602777 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.616067 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.620589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.620627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.620639 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.620655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.620664 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.635679 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: E0312 18:04:28.635823 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.637282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.637341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.637351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.637366 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.637375 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.739843 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.740314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.740325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.740346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.740358 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.841993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.842022 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.842030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.842042 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.842050 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.933345 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d37aa11-8fa5-4eb3-8edd-6f71523623b5" containerID="73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101" exitCode=0 Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.933428 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerDied","Data":"73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.939710 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.939742 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.939754 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.939763 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.939773 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.939781 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.944591 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.944663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.944675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.944713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.944726 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:28Z","lastTransitionTime":"2026-03-12T18:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.946189 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.960142 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.976246 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:28 crc kubenswrapper[4926]: I0312 18:04:28.994533 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.006976 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.021779 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.035519 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.049094 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.050550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.050600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.050615 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.050634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.050645 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.069882 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.087326 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.100706 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.153288 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.153335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.153349 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.153368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.153380 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.256484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.256550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.256568 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.256592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.256610 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.359118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.359177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.359193 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.359218 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.359236 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.462095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.462159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.462183 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.462212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.462235 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.502810 4926 scope.go:117] "RemoveContainer" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.502906 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.565590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.565637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.565649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.565664 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.565676 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.668144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.668190 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.668200 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.668213 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.668222 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.770734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.770793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.770815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.770846 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.770869 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.873835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.873902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.873925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.873957 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.873985 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.946375 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.950832 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.951021 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.956777 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d37aa11-8fa5-4eb3-8edd-6f71523623b5" containerID="099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e" exitCode=0 Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.956836 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerDied","Data":"099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.970635 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.982933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.983020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.983044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.983074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.983099 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:29Z","lastTransitionTime":"2026-03-12T18:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.984379 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-4gmrt"] Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.984794 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.987224 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:29Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.987545 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.987787 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.987946 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 12 18:04:29 crc kubenswrapper[4926]: I0312 18:04:29.998391 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.005181 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.028239 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.048639 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.072086 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.095670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.095705 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.095715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.095730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.095740 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.104312 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.122822 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.131494 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.131637 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-serviceca\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.131668 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:04:34.131645455 +0000 UTC m=+114.500271788 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.131708 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.131737 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-host\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.131765 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.131792 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmvtd\" (UniqueName: \"kubernetes.io/projected/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-kube-api-access-fmvtd\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.131830 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.131882 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.131915 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:34.131899572 +0000 UTC m=+114.500525905 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.131929 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:34.131922933 +0000 UTC m=+114.500549266 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.142584 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.155617 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.171288 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.184200 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.194293 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.198004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.198032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.198041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.198056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.198066 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.207212 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.220497 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.230094 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.230947 4926 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.232441 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.232498 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-serviceca\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.232536 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232542 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232560 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.232562 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-host\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232570 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.232594 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmvtd\" (UniqueName: \"kubernetes.io/projected/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-kube-api-access-fmvtd\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232607 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:34.232595914 +0000 UTC m=+114.601222247 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.232683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-host\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232728 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232757 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232770 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.232828 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:34.23280676 +0000 UTC m=+114.601433083 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.233533 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-serviceca\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.242561 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.261292 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.262879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmvtd\" (UniqueName: \"kubernetes.io/projected/dcfdbe34-faf7-4306-a2d8-6e95715f4f2a-kube-api-access-fmvtd\") pod \"node-ca-4gmrt\" (UID: \"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\") " pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.280799 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.300316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.300347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.300356 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.300370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.300380 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.304627 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.314988 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4gmrt" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.318116 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.329975 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: W0312 18:04:30.334992 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcfdbe34_faf7_4306_a2d8_6e95715f4f2a.slice/crio-a534e02499142e9db4e8c9bd3e0a002d9b687e211c85fa0d45f09ee3e7adb767 WatchSource:0}: Error finding container a534e02499142e9db4e8c9bd3e0a002d9b687e211c85fa0d45f09ee3e7adb767: Status 404 returned error can't find the container with id a534e02499142e9db4e8c9bd3e0a002d9b687e211c85fa0d45f09ee3e7adb767 Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.342396 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.354051 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.366050 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.402038 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.402113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.402127 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.402143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.402155 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.489665 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.489810 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.490044 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.490143 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.490227 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:30 crc kubenswrapper[4926]: E0312 18:04:30.490369 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.504004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.504051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.504068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.504089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.504104 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.506059 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.519961 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.534291 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.548360 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.567125 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.587074 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.605941 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.605991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.606001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.606018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.606030 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.608824 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.628616 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.645021 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.656925 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.668525 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.677403 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.684934 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.707926 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.707970 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.707981 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.707995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.708006 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.810495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.810545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.810562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.810585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.810601 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.913432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.913493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.913504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.913521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.913532 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:30Z","lastTransitionTime":"2026-03-12T18:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.963726 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d37aa11-8fa5-4eb3-8edd-6f71523623b5" containerID="5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2" exitCode=0 Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.963801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerDied","Data":"5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.965801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4gmrt" event={"ID":"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a","Type":"ContainerStarted","Data":"aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.965854 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4gmrt" event={"ID":"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a","Type":"ContainerStarted","Data":"a534e02499142e9db4e8c9bd3e0a002d9b687e211c85fa0d45f09ee3e7adb767"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.967944 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359"} Mar 12 18:04:30 crc kubenswrapper[4926]: I0312 18:04:30.983510 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.004063 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.016158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.016236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.016258 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.016281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.016298 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.018571 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.032394 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.045127 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.055753 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.075407 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.088614 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.100199 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.115036 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.119561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.119614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.119624 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.119640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.119650 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.128332 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.144777 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.157889 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.170763 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.182134 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.203550 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.215737 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.221596 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.221625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.221636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.221650 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.221662 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.231000 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.242424 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.253717 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.266123 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.276208 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.287252 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.298244 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.313419 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.323987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.324016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.324024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.324037 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.324045 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.328311 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.426338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.426394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.426412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.426445 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.426510 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.528913 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.528969 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.528986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.529009 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.529026 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.631367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.631477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.631495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.631518 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.631535 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.734325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.734365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.734376 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.734390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.734400 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.837212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.837629 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.837650 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.837676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.837693 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.940502 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.940571 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.940589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.940613 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.940629 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:31Z","lastTransitionTime":"2026-03-12T18:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.973210 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d37aa11-8fa5-4eb3-8edd-6f71523623b5" containerID="4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a" exitCode=0 Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.973316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerDied","Data":"4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.978547 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} Mar 12 18:04:31 crc kubenswrapper[4926]: I0312 18:04:31.991585 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:31Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.009587 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.032926 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.043019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.043056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.043066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.043082 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.043092 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.051847 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.062386 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.076792 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.089275 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.101344 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.111432 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.122191 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.133833 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.145128 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.146053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.146089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.146099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.146113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.146122 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.157811 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:32Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.248122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.248163 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.248173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.248188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.248199 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.351055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.351511 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.351751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.351947 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.352126 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.455605 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.455663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.455682 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.455710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.455726 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.489638 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.489679 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.489713 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:32 crc kubenswrapper[4926]: E0312 18:04:32.489789 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:32 crc kubenswrapper[4926]: E0312 18:04:32.489927 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:32 crc kubenswrapper[4926]: E0312 18:04:32.490041 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.558144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.558188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.558198 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.558214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.558224 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.660914 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.660957 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.660969 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.660986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.660997 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.763762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.763800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.763811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.763827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.763839 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.869876 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.869928 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.869938 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.869963 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.869973 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.972598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.972683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.972708 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.972738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.972763 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:32Z","lastTransitionTime":"2026-03-12T18:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.987394 4926 generic.go:334] "Generic (PLEG): container finished" podID="1d37aa11-8fa5-4eb3-8edd-6f71523623b5" containerID="a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef" exitCode=0 Mar 12 18:04:32 crc kubenswrapper[4926]: I0312 18:04:32.987483 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerDied","Data":"a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.009463 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.030648 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.051651 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.067782 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.074870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.074921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.074934 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.074951 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.074964 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.080229 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.093364 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.102035 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.112995 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.121489 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.136618 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.149992 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.165521 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.177184 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:33Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.177252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.177287 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.177298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.177311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.177321 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.280003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.280051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.280064 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.280083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.280095 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.383467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.383513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.383527 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.383547 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.383559 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.486543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.486592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.486607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.486629 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.486644 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.590910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.590962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.590980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.591001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.591021 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.694170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.695008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.695026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.695085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.695103 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.797729 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.797803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.797821 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.797848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.797868 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.900272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.900328 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.900339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.900361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.900377 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:33Z","lastTransitionTime":"2026-03-12T18:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:33 crc kubenswrapper[4926]: I0312 18:04:33.995763 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" event={"ID":"1d37aa11-8fa5-4eb3-8edd-6f71523623b5","Type":"ContainerStarted","Data":"eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.000573 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.000869 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.000912 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.000923 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.002027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.002063 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.002072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.002088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.002096 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.011657 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.029373 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.042545 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.094998 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.095073 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.095288 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.107824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.107866 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.107875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.107889 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.107898 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.109174 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.122288 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.132955 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.143567 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.151391 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.161348 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.170412 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.173731 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.173855 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:04:42.173831671 +0000 UTC m=+122.542458024 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.174185 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.174324 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:42.174310116 +0000 UTC m=+122.542936509 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.174324 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.174629 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.174736 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.174788 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:42.17477204 +0000 UTC m=+122.543398443 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.181330 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.193028 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.210555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.210593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.210602 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.210616 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.210627 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.212138 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.232997 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.244092 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.254705 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.272710 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.275166 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.275206 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275306 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275326 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275337 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275367 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275389 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275402 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275377 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:42.275365338 +0000 UTC m=+122.643991671 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.275481 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:42.275467241 +0000 UTC m=+122.644093574 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.286258 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.297781 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.311688 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.312681 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.312722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.312733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.312750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.312760 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.324994 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.338198 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.350537 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.360656 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.370086 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:34Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.415200 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.415242 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.415251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.415265 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.415275 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.489020 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.489133 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.489037 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.489257 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.489398 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:34 crc kubenswrapper[4926]: E0312 18:04:34.489465 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.517703 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.517766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.517785 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.517809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.517826 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.620282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.620338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.620363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.620393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.620409 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.724003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.724098 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.724115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.724138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.724155 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.827891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.827999 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.828016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.828045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.828064 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.931356 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.931500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.931522 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.931550 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:34 crc kubenswrapper[4926]: I0312 18:04:34.931570 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:34Z","lastTransitionTime":"2026-03-12T18:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.033186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.033240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.033251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.033268 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.033284 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.136262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.136353 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.136369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.136392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.136407 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.239243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.239286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.239321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.239338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.239349 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.341971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.342041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.342063 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.342091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.342112 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.445339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.445410 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.445433 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.445498 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.445522 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.549127 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.549194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.549220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.549250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.549274 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.652207 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.652259 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.652278 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.652302 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.652319 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.755820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.755878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.755899 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.755925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.755945 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.858850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.858883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.858894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.858910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.858921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.961221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.961252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.961262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.961279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.961292 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:35Z","lastTransitionTime":"2026-03-12T18:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.993103 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc"] Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.993639 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.997013 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 12 18:04:35 crc kubenswrapper[4926]: I0312 18:04:35.997062 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.013932 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.032390 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.047970 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.063503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.063723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.063801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.063960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.064066 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.064145 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.082746 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.095507 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12de8a94-72e6-4d72-8e39-42f3ef9d1125-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.095575 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12de8a94-72e6-4d72-8e39-42f3ef9d1125-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.095599 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12de8a94-72e6-4d72-8e39-42f3ef9d1125-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.095695 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwn2v\" (UniqueName: \"kubernetes.io/projected/12de8a94-72e6-4d72-8e39-42f3ef9d1125-kube-api-access-vwn2v\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.095879 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.107892 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.118621 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.129654 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.140675 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.151407 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.166484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.166536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.166549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.166565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.166577 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.169075 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.179704 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.193809 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.196209 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12de8a94-72e6-4d72-8e39-42f3ef9d1125-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.196270 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12de8a94-72e6-4d72-8e39-42f3ef9d1125-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.196290 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12de8a94-72e6-4d72-8e39-42f3ef9d1125-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.196326 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwn2v\" (UniqueName: \"kubernetes.io/projected/12de8a94-72e6-4d72-8e39-42f3ef9d1125-kube-api-access-vwn2v\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.197068 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12de8a94-72e6-4d72-8e39-42f3ef9d1125-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.197194 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12de8a94-72e6-4d72-8e39-42f3ef9d1125-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.202269 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12de8a94-72e6-4d72-8e39-42f3ef9d1125-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.217269 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwn2v\" (UniqueName: \"kubernetes.io/projected/12de8a94-72e6-4d72-8e39-42f3ef9d1125-kube-api-access-vwn2v\") pod \"ovnkube-control-plane-749d76644c-fq9dc\" (UID: \"12de8a94-72e6-4d72-8e39-42f3ef9d1125\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.275257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.275298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.275313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.275336 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.275353 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.309375 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" Mar 12 18:04:36 crc kubenswrapper[4926]: W0312 18:04:36.325021 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12de8a94_72e6_4d72_8e39_42f3ef9d1125.slice/crio-f84fc990f02e81fbca9e4f4d9a207c3bbefbfb7135fc031b4ce43b6a9f2d9125 WatchSource:0}: Error finding container f84fc990f02e81fbca9e4f4d9a207c3bbefbfb7135fc031b4ce43b6a9f2d9125: Status 404 returned error can't find the container with id f84fc990f02e81fbca9e4f4d9a207c3bbefbfb7135fc031b4ce43b6a9f2d9125 Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.383845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.383907 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.383919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.383939 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.383951 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.486596 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.486651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.486712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.486735 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.486748 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.488871 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.488893 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.488867 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:36 crc kubenswrapper[4926]: E0312 18:04:36.488999 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:36 crc kubenswrapper[4926]: E0312 18:04:36.489077 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:36 crc kubenswrapper[4926]: E0312 18:04:36.489208 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.589475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.589869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.589881 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.589903 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.589914 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.693024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.693088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.693109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.693134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.693152 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.714656 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-n7pd7"] Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.715577 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:36 crc kubenswrapper[4926]: E0312 18:04:36.715707 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.733808 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.753962 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.783155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.795363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.795612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.795697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.795818 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.795910 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.800242 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.801683 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6bnm\" (UniqueName: \"kubernetes.io/projected/211eeae6-9b41-484b-bd13-99c1c28cdf96-kube-api-access-n6bnm\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.801728 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.809310 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.821953 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.836737 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.847070 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.860099 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.880582 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.898627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.898675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.898713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.898731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.898743 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:36Z","lastTransitionTime":"2026-03-12T18:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.900107 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.902692 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6bnm\" (UniqueName: \"kubernetes.io/projected/211eeae6-9b41-484b-bd13-99c1c28cdf96-kube-api-access-n6bnm\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.902789 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:36 crc kubenswrapper[4926]: E0312 18:04:36.902949 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:36 crc kubenswrapper[4926]: E0312 18:04:36.903029 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:37.403005018 +0000 UTC m=+117.771631391 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.917689 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.932630 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.940697 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6bnm\" (UniqueName: \"kubernetes.io/projected/211eeae6-9b41-484b-bd13-99c1c28cdf96-kube-api-access-n6bnm\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.944713 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:36 crc kubenswrapper[4926]: I0312 18:04:36.954600 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:36Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.001371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.001721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.001831 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.001948 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.002211 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.010843 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" event={"ID":"12de8a94-72e6-4d72-8e39-42f3ef9d1125","Type":"ContainerStarted","Data":"f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.010881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" event={"ID":"12de8a94-72e6-4d72-8e39-42f3ef9d1125","Type":"ContainerStarted","Data":"9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.010892 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" event={"ID":"12de8a94-72e6-4d72-8e39-42f3ef9d1125","Type":"ContainerStarted","Data":"f84fc990f02e81fbca9e4f4d9a207c3bbefbfb7135fc031b4ce43b6a9f2d9125"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.013598 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/0.log" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.017992 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115" exitCode=1 Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.018053 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.019123 4926 scope.go:117] "RemoveContainer" containerID="be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.030537 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.043785 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.067578 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.083369 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.097281 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.105746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.105788 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.105800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.105816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.105827 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.123468 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.138967 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.150476 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.165655 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.177369 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.187059 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.203849 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.207529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.207551 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.207560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.207575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.207586 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.219765 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.232088 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.244036 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.256696 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.269298 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.282593 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.295059 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.305144 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.309837 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.309874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.309888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.309904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.309915 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.317655 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.329879 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.347049 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"I0312 18:04:36.449007 6832 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 18:04:36.449037 6832 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 18:04:36.449512 6832 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 18:04:36.449524 6832 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 18:04:36.449541 6832 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 18:04:36.449629 6832 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:36.449657 6832 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:36.449688 6832 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:36.449696 6832 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:36.449714 6832 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 18:04:36.449719 6832 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:36.449738 6832 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:36.449746 6832 factory.go:656] Stopping watch factory\\\\nI0312 18:04:36.449751 6832 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:36.449761 6832 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:36.449740 6832 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.361122 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.374909 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.386990 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.397112 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.404882 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.411672 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:37 crc kubenswrapper[4926]: E0312 18:04:37.411793 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:37 crc kubenswrapper[4926]: E0312 18:04:37.411843 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:38.411828051 +0000 UTC m=+118.780454384 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.412789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.412833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.412842 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.412861 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.412873 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.416944 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.432890 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:37Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.514847 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.514902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.514914 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.514932 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.514943 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.627646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.627688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.627699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.627716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.627728 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.730918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.730989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.731006 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.731034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.731051 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.833565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.833604 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.833621 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.833636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.833646 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.936282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.936317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.936325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.936341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:37 crc kubenswrapper[4926]: I0312 18:04:37.936350 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:37Z","lastTransitionTime":"2026-03-12T18:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.024171 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/0.log" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.028861 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.029474 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.039466 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.039507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.039521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.039541 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.039557 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.043584 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.062553 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.080560 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.099824 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"I0312 18:04:36.449007 6832 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 18:04:36.449037 6832 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 18:04:36.449512 6832 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 18:04:36.449524 6832 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 18:04:36.449541 6832 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 18:04:36.449629 6832 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:36.449657 6832 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:36.449688 6832 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:36.449696 6832 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:36.449714 6832 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 18:04:36.449719 6832 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:36.449738 6832 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:36.449746 6832 factory.go:656] Stopping watch factory\\\\nI0312 18:04:36.449751 6832 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:36.449761 6832 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:36.449740 6832 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.121814 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.135940 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.141177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.141228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.141239 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.141255 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.141266 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.153196 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.166685 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.178850 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.193125 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.203068 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.215361 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.225675 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.238647 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.244025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.244058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.244071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.244086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.244099 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.249774 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.346339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.346392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.346404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.346421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.346432 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.432583 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.432753 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.432866 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:40.432838124 +0000 UTC m=+120.801464497 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.450257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.450326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.450345 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.450372 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.450391 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.489800 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.490010 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.490046 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.490071 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.490142 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.490283 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.490415 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.490533 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.502878 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.553580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.553656 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.553667 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.553681 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.553691 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.652212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.652304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.652325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.652349 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.652364 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.671267 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.681257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.681394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.681432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.681523 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.681544 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.702652 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.707396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.707510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.707535 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.707566 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.707588 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.727639 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.731766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.731856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.731883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.731898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.731911 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.744686 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.748814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.748840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.748851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.748866 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.748899 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.762564 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:38Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:38 crc kubenswrapper[4926]: E0312 18:04:38.762714 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.763962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.763995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.764004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.764020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.764030 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.866663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.866709 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.866720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.866738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.866749 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.970214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.970247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.970256 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.970272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:38 crc kubenswrapper[4926]: I0312 18:04:38.970284 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:38Z","lastTransitionTime":"2026-03-12T18:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.034891 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/1.log" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.035788 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/0.log" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.038938 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572" exitCode=1 Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.039036 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.039109 4926 scope.go:117] "RemoveContainer" containerID="be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.041065 4926 scope.go:117] "RemoveContainer" containerID="3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572" Mar 12 18:04:39 crc kubenswrapper[4926]: E0312 18:04:39.041289 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.064815 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.072565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.072606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.072618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.072635 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.072649 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.089765 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.105439 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.118328 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.135717 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.153122 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.169118 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.174371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.174400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.174411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.174431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.174462 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.188863 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be6eea7ebb0023aaf587a20ffee0cf7dd08281f2265ae735dcb8e3486dfc2115\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"I0312 18:04:36.449007 6832 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 18:04:36.449037 6832 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 18:04:36.449512 6832 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 18:04:36.449524 6832 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 18:04:36.449541 6832 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 18:04:36.449629 6832 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:36.449657 6832 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:36.449688 6832 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:36.449696 6832 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:36.449714 6832 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 18:04:36.449719 6832 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:36.449738 6832 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:36.449746 6832 factory.go:656] Stopping watch factory\\\\nI0312 18:04:36.449751 6832 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:36.449761 6832 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:36.449740 6832 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"ft/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.056752 7017 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.057188 7017 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:38.057288 7017 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:38.057368 7017 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:38.057393 7017 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:38.057474 7017 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:38.057564 7017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:38.057640 7017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:38.057709 7017 factory.go:656] Stopping watch factory\\\\nI0312 18:04:38.057731 7017 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:38.057773 7017 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 18:04:38.057798 7017 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:38.057895 7017 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.203157 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.212664 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.227188 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.238817 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.254119 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.264226 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.277000 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.277153 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.277185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.277199 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.277217 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.277229 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.287782 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:39Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.379912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.379973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.379990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.380014 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.380031 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.482751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.482798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.482813 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.482828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.482839 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.585146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.585201 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.585214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.585231 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.585242 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.689016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.689115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.689135 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.689588 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.689811 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.792936 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.793013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.793035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.793062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.793079 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.895206 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.895253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.895263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.895275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.895285 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.998577 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.998634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.998654 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.998677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:39 crc kubenswrapper[4926]: I0312 18:04:39.998693 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:39Z","lastTransitionTime":"2026-03-12T18:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.045435 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/1.log" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.050702 4926 scope.go:117] "RemoveContainer" containerID="3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.050866 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.073765 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.095028 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.103350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.103379 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.103388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.103401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.103410 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:40Z","lastTransitionTime":"2026-03-12T18:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.115215 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.132970 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.155134 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.170509 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.185991 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.202246 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.206471 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.206518 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.206535 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.206558 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.206579 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:40Z","lastTransitionTime":"2026-03-12T18:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.222656 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.291872 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"ft/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.056752 7017 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.057188 7017 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:38.057288 7017 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:38.057368 7017 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:38.057393 7017 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:38.057474 7017 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:38.057564 7017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:38.057640 7017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:38.057709 7017 factory.go:656] Stopping watch factory\\\\nI0312 18:04:38.057731 7017 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:38.057773 7017 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 18:04:38.057798 7017 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:38.057895 7017 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.306364 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.309484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.309521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.309533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.309549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.309560 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:40Z","lastTransitionTime":"2026-03-12T18:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.323146 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.336804 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.350167 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.361719 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.373944 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.412212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.412256 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.412265 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.412278 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.412305 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:40Z","lastTransitionTime":"2026-03-12T18:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.476068 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.476219 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.476284 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:44.476266905 +0000 UTC m=+124.844893258 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.492507 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.492657 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.493042 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.493125 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.493188 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.493250 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.493308 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.493380 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.512482 4926 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.512584 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.529097 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.544403 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.556736 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.573288 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: E0312 18:04:40.577752 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.589832 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.606160 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.625148 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"ft/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.056752 7017 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.057188 7017 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:38.057288 7017 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:38.057368 7017 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:38.057393 7017 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:38.057474 7017 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:38.057564 7017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:38.057640 7017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:38.057709 7017 factory.go:656] Stopping watch factory\\\\nI0312 18:04:38.057731 7017 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:38.057773 7017 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 18:04:38.057798 7017 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:38.057895 7017 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.642691 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.655058 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.669906 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.683737 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.695858 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.705378 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.715879 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:40 crc kubenswrapper[4926]: I0312 18:04:40.726560 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.234894 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.250358 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.264339 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.279534 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.297485 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.312923 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.336089 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"ft/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.056752 7017 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.057188 7017 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:38.057288 7017 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:38.057368 7017 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:38.057393 7017 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:38.057474 7017 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:38.057564 7017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:38.057640 7017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:38.057709 7017 factory.go:656] Stopping watch factory\\\\nI0312 18:04:38.057731 7017 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:38.057773 7017 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 18:04:38.057798 7017 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:38.057895 7017 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.374253 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.386112 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.400391 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.414290 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.444510 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.465265 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.485878 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.501027 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.514291 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:41 crc kubenswrapper[4926]: I0312 18:04:41.528762 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.188650 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.188877 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:04:58.188839358 +0000 UTC m=+138.557465731 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.189121 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.189232 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.189377 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.189543 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:58.18952182 +0000 UTC m=+138.558148233 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.189396 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.189740 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:58.189711565 +0000 UTC m=+138.558337978 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.290751 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.290894 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.290986 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.291023 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.291036 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.291087 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.291106 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:58.291084878 +0000 UTC m=+138.659711211 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.291116 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.291136 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.291225 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:58.291185481 +0000 UTC m=+138.659811904 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.489608 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.490060 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.489695 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.489731 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:42 crc kubenswrapper[4926]: I0312 18:04:42.489671 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.490587 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.490328 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:42 crc kubenswrapper[4926]: E0312 18:04:42.490714 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:44 crc kubenswrapper[4926]: I0312 18:04:44.489506 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:44 crc kubenswrapper[4926]: I0312 18:04:44.489533 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:44 crc kubenswrapper[4926]: E0312 18:04:44.489671 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:44 crc kubenswrapper[4926]: E0312 18:04:44.489798 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:44 crc kubenswrapper[4926]: I0312 18:04:44.489596 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:44 crc kubenswrapper[4926]: I0312 18:04:44.489905 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:44 crc kubenswrapper[4926]: E0312 18:04:44.489953 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:44 crc kubenswrapper[4926]: E0312 18:04:44.490062 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:44 crc kubenswrapper[4926]: I0312 18:04:44.514997 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:44 crc kubenswrapper[4926]: E0312 18:04:44.515201 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:44 crc kubenswrapper[4926]: E0312 18:04:44.515280 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:04:52.515256662 +0000 UTC m=+132.883883035 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:45 crc kubenswrapper[4926]: E0312 18:04:45.579138 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:04:46 crc kubenswrapper[4926]: I0312 18:04:46.489620 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:46 crc kubenswrapper[4926]: I0312 18:04:46.489689 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:46 crc kubenswrapper[4926]: E0312 18:04:46.490186 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:46 crc kubenswrapper[4926]: I0312 18:04:46.489784 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:46 crc kubenswrapper[4926]: I0312 18:04:46.489746 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:46 crc kubenswrapper[4926]: E0312 18:04:46.490621 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:46 crc kubenswrapper[4926]: E0312 18:04:46.490771 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:46 crc kubenswrapper[4926]: E0312 18:04:46.490832 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:48 crc kubenswrapper[4926]: I0312 18:04:48.489874 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:48 crc kubenswrapper[4926]: I0312 18:04:48.489893 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:48 crc kubenswrapper[4926]: I0312 18:04:48.490010 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:48 crc kubenswrapper[4926]: I0312 18:04:48.489967 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:48 crc kubenswrapper[4926]: E0312 18:04:48.490144 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:48 crc kubenswrapper[4926]: E0312 18:04:48.490304 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:48 crc kubenswrapper[4926]: E0312 18:04:48.490420 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:48 crc kubenswrapper[4926]: E0312 18:04:48.490599 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.134292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.134353 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.134371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.134394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.134411 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:49Z","lastTransitionTime":"2026-03-12T18:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:49 crc kubenswrapper[4926]: E0312 18:04:49.150244 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:49Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.157753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.157786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.157795 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.157808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.157816 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:49Z","lastTransitionTime":"2026-03-12T18:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:49 crc kubenswrapper[4926]: E0312 18:04:49.178023 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:49Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.182848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.182887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.182900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.182916 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.182926 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:49Z","lastTransitionTime":"2026-03-12T18:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:49 crc kubenswrapper[4926]: E0312 18:04:49.198483 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:49Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.203153 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.203205 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.203215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.203235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.203250 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:49Z","lastTransitionTime":"2026-03-12T18:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:49 crc kubenswrapper[4926]: E0312 18:04:49.219599 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:49Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.223369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.223426 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.223467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.223492 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:49 crc kubenswrapper[4926]: I0312 18:04:49.223509 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:49Z","lastTransitionTime":"2026-03-12T18:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:49 crc kubenswrapper[4926]: E0312 18:04:49.237005 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:49Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:49 crc kubenswrapper[4926]: E0312 18:04:49.237200 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.489189 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.489253 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.489227 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.489204 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:50 crc kubenswrapper[4926]: E0312 18:04:50.489350 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:50 crc kubenswrapper[4926]: E0312 18:04:50.489521 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:50 crc kubenswrapper[4926]: E0312 18:04:50.489758 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:50 crc kubenswrapper[4926]: E0312 18:04:50.489824 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.500594 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.515533 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.529985 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.544020 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.558521 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.569343 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: E0312 18:04:50.580070 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.582260 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.599957 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.621746 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"ft/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.056752 7017 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.057188 7017 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:38.057288 7017 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:38.057368 7017 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:38.057393 7017 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:38.057474 7017 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:38.057564 7017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:38.057640 7017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:38.057709 7017 factory.go:656] Stopping watch factory\\\\nI0312 18:04:38.057731 7017 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:38.057773 7017 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 18:04:38.057798 7017 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:38.057895 7017 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.637488 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.649228 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.666344 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.682252 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.701410 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.718479 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:50 crc kubenswrapper[4926]: I0312 18:04:50.738899 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:50Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:52 crc kubenswrapper[4926]: I0312 18:04:52.489624 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:52 crc kubenswrapper[4926]: I0312 18:04:52.489683 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:52 crc kubenswrapper[4926]: E0312 18:04:52.489795 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:52 crc kubenswrapper[4926]: I0312 18:04:52.489908 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:52 crc kubenswrapper[4926]: I0312 18:04:52.489971 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:52 crc kubenswrapper[4926]: E0312 18:04:52.490271 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:52 crc kubenswrapper[4926]: E0312 18:04:52.490401 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:52 crc kubenswrapper[4926]: E0312 18:04:52.490579 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:52 crc kubenswrapper[4926]: I0312 18:04:52.600660 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:52 crc kubenswrapper[4926]: E0312 18:04:52.600840 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:52 crc kubenswrapper[4926]: E0312 18:04:52.600906 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:05:08.600884861 +0000 UTC m=+148.969511224 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:04:54 crc kubenswrapper[4926]: I0312 18:04:54.490006 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:54 crc kubenswrapper[4926]: I0312 18:04:54.490056 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:54 crc kubenswrapper[4926]: I0312 18:04:54.490104 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:54 crc kubenswrapper[4926]: I0312 18:04:54.490056 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:54 crc kubenswrapper[4926]: E0312 18:04:54.490245 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:54 crc kubenswrapper[4926]: E0312 18:04:54.490397 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:54 crc kubenswrapper[4926]: E0312 18:04:54.490576 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:54 crc kubenswrapper[4926]: E0312 18:04:54.490732 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:55 crc kubenswrapper[4926]: I0312 18:04:55.490849 4926 scope.go:117] "RemoveContainer" containerID="3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572" Mar 12 18:04:55 crc kubenswrapper[4926]: E0312 18:04:55.581783 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.108169 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/1.log" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.111111 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd"} Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.111479 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.122717 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.140131 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.156324 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.168615 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.185553 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.197793 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.215377 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.231110 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.245252 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.269573 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.289116 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.311672 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"ft/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.056752 7017 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.057188 7017 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:38.057288 7017 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:38.057368 7017 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:38.057393 7017 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:38.057474 7017 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:38.057564 7017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:38.057640 7017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:38.057709 7017 factory.go:656] Stopping watch factory\\\\nI0312 18:04:38.057731 7017 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:38.057773 7017 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 18:04:38.057798 7017 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:38.057895 7017 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.359896 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.373558 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.385808 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.397410 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:56Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.489802 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.489857 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.489820 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:56 crc kubenswrapper[4926]: E0312 18:04:56.489919 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:56 crc kubenswrapper[4926]: E0312 18:04:56.490097 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:56 crc kubenswrapper[4926]: I0312 18:04:56.490186 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:56 crc kubenswrapper[4926]: E0312 18:04:56.490269 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:56 crc kubenswrapper[4926]: E0312 18:04:56.490353 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.117506 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/2.log" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.119007 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/1.log" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.123716 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd" exitCode=1 Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.123817 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd"} Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.123984 4926 scope.go:117] "RemoveContainer" containerID="3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.125071 4926 scope.go:117] "RemoveContainer" containerID="fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd" Mar 12 18:04:57 crc kubenswrapper[4926]: E0312 18:04:57.125387 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.145823 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.163393 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.181487 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.197293 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.211702 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.227732 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.241364 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.272425 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e57e6a3873367656206dd5f7b6b7302dd876f3735f7e7fc4e741cd61f966572\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:38Z\\\",\\\"message\\\":\\\"ft/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.056752 7017 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 18:04:38.057188 7017 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:38.057288 7017 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:38.057368 7017 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 18:04:38.057393 7017 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:38.057474 7017 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:38.057564 7017 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 18:04:38.057640 7017 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 18:04:38.057709 7017 factory.go:656] Stopping watch factory\\\\nI0312 18:04:38.057731 7017 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:38.057773 7017 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 18:04:38.057798 7017 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:38.057895 7017 ovnkube.go:137] failed to run ovnkube: [failed to start network cont\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.296536 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.312034 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.327249 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.340763 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.354736 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.365682 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.383122 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:57 crc kubenswrapper[4926]: I0312 18:04:57.397606 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:57Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.129815 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/2.log" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.134019 4926 scope.go:117] "RemoveContainer" containerID="fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.134266 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.154282 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.170716 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.189414 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.205056 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.217249 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.231832 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.250242 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.263759 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.271522 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.271938 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:05:30.271901526 +0000 UTC m=+170.640527869 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.272086 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.273731 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.273851 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.273730 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.273930 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:05:30.273884505 +0000 UTC m=+170.642510878 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.274129 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:05:30.274084411 +0000 UTC m=+170.642710774 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.301730 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.323299 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.343002 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.359399 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.373065 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.375452 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.375589 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.375856 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.375951 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.376029 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.376144 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:05:30.376126794 +0000 UTC m=+170.744753127 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.376279 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.376311 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.378547 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.378683 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:05:30.378655379 +0000 UTC m=+170.747281752 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.385061 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.397289 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.407931 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:58Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.488964 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.489032 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.488993 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.489192 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.489329 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.489515 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:04:58 crc kubenswrapper[4926]: I0312 18:04:58.489905 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:04:58 crc kubenswrapper[4926]: E0312 18:04:58.490174 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.549381 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.549467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.549488 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.549516 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.549535 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:59Z","lastTransitionTime":"2026-03-12T18:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:59 crc kubenswrapper[4926]: E0312 18:04:59.569760 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:59Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.574483 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.574553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.574579 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.574611 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.574637 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:59Z","lastTransitionTime":"2026-03-12T18:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:59 crc kubenswrapper[4926]: E0312 18:04:59.595518 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:59Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.601346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.601411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.601435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.601498 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.601520 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:59Z","lastTransitionTime":"2026-03-12T18:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:59 crc kubenswrapper[4926]: E0312 18:04:59.618326 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:59Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.623168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.623223 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.623241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.623263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.623280 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:59Z","lastTransitionTime":"2026-03-12T18:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:59 crc kubenswrapper[4926]: E0312 18:04:59.637957 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:59Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.641912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.641967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.641984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.642007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:04:59 crc kubenswrapper[4926]: I0312 18:04:59.642024 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:04:59Z","lastTransitionTime":"2026-03-12T18:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:04:59 crc kubenswrapper[4926]: E0312 18:04:59.656927 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:04:59Z is after 2025-08-24T17:21:41Z" Mar 12 18:04:59 crc kubenswrapper[4926]: E0312 18:04:59.657161 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.489574 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.489698 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.489603 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:00 crc kubenswrapper[4926]: E0312 18:05:00.489813 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.489603 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:00 crc kubenswrapper[4926]: E0312 18:05:00.489945 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:00 crc kubenswrapper[4926]: E0312 18:05:00.490031 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:00 crc kubenswrapper[4926]: E0312 18:05:00.490131 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.505468 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.524303 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.543274 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.562911 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: E0312 18:05:00.583198 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.583167 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.605504 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.628182 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.643956 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.657974 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.671087 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.685288 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.696085 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.709898 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.722822 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.739959 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:00 crc kubenswrapper[4926]: I0312 18:05:00.759731 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:00Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:02 crc kubenswrapper[4926]: I0312 18:05:02.490067 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:02 crc kubenswrapper[4926]: E0312 18:05:02.491275 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:02 crc kubenswrapper[4926]: I0312 18:05:02.490171 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:02 crc kubenswrapper[4926]: E0312 18:05:02.491775 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:02 crc kubenswrapper[4926]: I0312 18:05:02.490122 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:02 crc kubenswrapper[4926]: I0312 18:05:02.490278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:02 crc kubenswrapper[4926]: E0312 18:05:02.492176 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:02 crc kubenswrapper[4926]: E0312 18:05:02.492324 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:04 crc kubenswrapper[4926]: I0312 18:05:04.489330 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:04 crc kubenswrapper[4926]: E0312 18:05:04.489516 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:04 crc kubenswrapper[4926]: I0312 18:05:04.489578 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:04 crc kubenswrapper[4926]: E0312 18:05:04.489816 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:04 crc kubenswrapper[4926]: I0312 18:05:04.489863 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:04 crc kubenswrapper[4926]: E0312 18:05:04.489948 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:04 crc kubenswrapper[4926]: I0312 18:05:04.490582 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:04 crc kubenswrapper[4926]: E0312 18:05:04.490664 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:05 crc kubenswrapper[4926]: E0312 18:05:05.584472 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:06 crc kubenswrapper[4926]: I0312 18:05:06.489988 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:06 crc kubenswrapper[4926]: I0312 18:05:06.490015 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:06 crc kubenswrapper[4926]: I0312 18:05:06.490245 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:06 crc kubenswrapper[4926]: I0312 18:05:06.490264 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:06 crc kubenswrapper[4926]: E0312 18:05:06.491086 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:06 crc kubenswrapper[4926]: E0312 18:05:06.491628 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:06 crc kubenswrapper[4926]: E0312 18:05:06.491853 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:06 crc kubenswrapper[4926]: E0312 18:05:06.492024 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:08 crc kubenswrapper[4926]: I0312 18:05:08.489266 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:08 crc kubenswrapper[4926]: I0312 18:05:08.489267 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:08 crc kubenswrapper[4926]: I0312 18:05:08.489376 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:08 crc kubenswrapper[4926]: E0312 18:05:08.489521 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:08 crc kubenswrapper[4926]: E0312 18:05:08.489668 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:08 crc kubenswrapper[4926]: E0312 18:05:08.489833 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:08 crc kubenswrapper[4926]: I0312 18:05:08.491150 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:08 crc kubenswrapper[4926]: E0312 18:05:08.491516 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:08 crc kubenswrapper[4926]: E0312 18:05:08.687695 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:05:08 crc kubenswrapper[4926]: E0312 18:05:08.687930 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:05:40.68781576 +0000 UTC m=+181.056442133 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:05:08 crc kubenswrapper[4926]: I0312 18:05:08.688568 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.039027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.039094 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.039112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.039137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.039155 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:10Z","lastTransitionTime":"2026-03-12T18:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.059579 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.064493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.064586 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.064618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.064655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.064694 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:10Z","lastTransitionTime":"2026-03-12T18:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.087877 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.093841 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.093894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.093911 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.093938 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.093955 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:10Z","lastTransitionTime":"2026-03-12T18:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.117353 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.122505 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.122578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.122603 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.122632 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.122655 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:10Z","lastTransitionTime":"2026-03-12T18:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.141428 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.145856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.145904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.145920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.145943 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.145960 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:10Z","lastTransitionTime":"2026-03-12T18:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.159836 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.159991 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.489680 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.489867 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.489908 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.489943 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.489985 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.490044 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.490523 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.490651 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.491245 4926 scope.go:117] "RemoveContainer" containerID="fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd" Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.491530 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.505671 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.532987 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.555223 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.568970 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: E0312 18:05:10.585715 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.589117 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.603299 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.614765 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.627776 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.642034 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.657501 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.675848 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.689737 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.705950 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.721477 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.748283 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:10 crc kubenswrapper[4926]: I0312 18:05:10.777249 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:10Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:12 crc kubenswrapper[4926]: I0312 18:05:12.488979 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:12 crc kubenswrapper[4926]: E0312 18:05:12.489195 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:12 crc kubenswrapper[4926]: I0312 18:05:12.489328 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:12 crc kubenswrapper[4926]: E0312 18:05:12.489521 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:12 crc kubenswrapper[4926]: I0312 18:05:12.489651 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:12 crc kubenswrapper[4926]: E0312 18:05:12.489788 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:12 crc kubenswrapper[4926]: I0312 18:05:12.489945 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:12 crc kubenswrapper[4926]: E0312 18:05:12.490199 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.191229 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/0.log" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.191311 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5a53ef4-c701-457f-9cf2-85819bf04d1a" containerID="54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c" exitCode=1 Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.191362 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerDied","Data":"54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c"} Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.192013 4926 scope.go:117] "RemoveContainer" containerID="54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.210773 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.231582 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.250040 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.273155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.286895 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.305691 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.322582 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.354087 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.376457 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.393499 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.412188 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.427136 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.441666 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.454018 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.465461 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.477801 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:14Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.489177 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.489307 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.489309 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:14 crc kubenswrapper[4926]: I0312 18:05:14.489215 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:14 crc kubenswrapper[4926]: E0312 18:05:14.489504 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:14 crc kubenswrapper[4926]: E0312 18:05:14.489651 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:14 crc kubenswrapper[4926]: E0312 18:05:14.489833 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:14 crc kubenswrapper[4926]: E0312 18:05:14.489988 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.197573 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/0.log" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.197650 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerStarted","Data":"bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4"} Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.214563 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.228783 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.244863 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.261409 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.275010 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.288975 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.302961 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.314807 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.329714 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.347795 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.367538 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.398188 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.421863 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.437920 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.455500 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: I0312 18:05:15.474897 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:15Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:15 crc kubenswrapper[4926]: E0312 18:05:15.587062 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:16 crc kubenswrapper[4926]: I0312 18:05:16.489747 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:16 crc kubenswrapper[4926]: I0312 18:05:16.489767 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:16 crc kubenswrapper[4926]: I0312 18:05:16.489750 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:16 crc kubenswrapper[4926]: I0312 18:05:16.489860 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:16 crc kubenswrapper[4926]: E0312 18:05:16.490068 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:16 crc kubenswrapper[4926]: E0312 18:05:16.490171 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:16 crc kubenswrapper[4926]: E0312 18:05:16.489967 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:16 crc kubenswrapper[4926]: E0312 18:05:16.490360 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:18 crc kubenswrapper[4926]: I0312 18:05:18.489517 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:18 crc kubenswrapper[4926]: I0312 18:05:18.489577 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:18 crc kubenswrapper[4926]: E0312 18:05:18.489735 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:18 crc kubenswrapper[4926]: I0312 18:05:18.489799 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:18 crc kubenswrapper[4926]: I0312 18:05:18.489518 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:18 crc kubenswrapper[4926]: E0312 18:05:18.489928 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:18 crc kubenswrapper[4926]: E0312 18:05:18.490080 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:18 crc kubenswrapper[4926]: E0312 18:05:18.490166 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.368329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.368382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.368395 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.368414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.368482 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:20Z","lastTransitionTime":"2026-03-12T18:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.381284 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.385017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.385066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.385079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.385096 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.385108 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:20Z","lastTransitionTime":"2026-03-12T18:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.400173 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.404122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.404264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.404374 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.404513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.404633 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:20Z","lastTransitionTime":"2026-03-12T18:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.418004 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.426759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.426807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.426827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.426844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.426855 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:20Z","lastTransitionTime":"2026-03-12T18:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.442954 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.448365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.448461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.448480 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.448504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.448522 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:20Z","lastTransitionTime":"2026-03-12T18:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.464313 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.464591 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.489550 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.489590 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.489651 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.489554 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.489782 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.489935 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.489996 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.490058 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.508912 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.525034 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.549098 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.566427 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.578391 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: E0312 18:05:20.587826 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.591352 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.606240 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.619334 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.633550 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.650257 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.669203 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.683920 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.699938 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.717854 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.734255 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:20 crc kubenswrapper[4926]: I0312 18:05:20.748570 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:20Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:22 crc kubenswrapper[4926]: I0312 18:05:22.489659 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:22 crc kubenswrapper[4926]: I0312 18:05:22.489683 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:22 crc kubenswrapper[4926]: I0312 18:05:22.489725 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:22 crc kubenswrapper[4926]: I0312 18:05:22.489744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:22 crc kubenswrapper[4926]: E0312 18:05:22.492084 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:22 crc kubenswrapper[4926]: E0312 18:05:22.492265 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:22 crc kubenswrapper[4926]: E0312 18:05:22.492635 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:22 crc kubenswrapper[4926]: E0312 18:05:22.492830 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:24 crc kubenswrapper[4926]: I0312 18:05:24.489345 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:24 crc kubenswrapper[4926]: I0312 18:05:24.489409 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:24 crc kubenswrapper[4926]: E0312 18:05:24.489579 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:24 crc kubenswrapper[4926]: E0312 18:05:24.489707 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:24 crc kubenswrapper[4926]: I0312 18:05:24.489804 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:24 crc kubenswrapper[4926]: E0312 18:05:24.489888 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:24 crc kubenswrapper[4926]: I0312 18:05:24.490065 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:24 crc kubenswrapper[4926]: E0312 18:05:24.490168 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:25 crc kubenswrapper[4926]: I0312 18:05:25.491206 4926 scope.go:117] "RemoveContainer" containerID="fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd" Mar 12 18:05:25 crc kubenswrapper[4926]: E0312 18:05:25.589013 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.237336 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/2.log" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.240254 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.240855 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.260345 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.275877 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.294376 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.310529 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.323984 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.338486 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.353750 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.365109 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.377392 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.403395 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.418160 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.428796 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.453112 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.468448 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.483343 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.489225 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.489280 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.489286 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:26 crc kubenswrapper[4926]: E0312 18:05:26.489350 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:26 crc kubenswrapper[4926]: E0312 18:05:26.489466 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.489493 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:26 crc kubenswrapper[4926]: E0312 18:05:26.489595 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:26 crc kubenswrapper[4926]: E0312 18:05:26.489769 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:26 crc kubenswrapper[4926]: I0312 18:05:26.498494 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:26Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.245661 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/3.log" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.246721 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/2.log" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.249761 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" exitCode=1 Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.249818 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.249866 4926 scope.go:117] "RemoveContainer" containerID="fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.250959 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:05:27 crc kubenswrapper[4926]: E0312 18:05:27.251481 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.277619 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fe14ac05f79c0314c60edfebe5cb903048be0cc438846a149ac872d84d8178cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:04:56Z\\\",\\\"message\\\":\\\"712973235162149816) with []\\\\nI0312 18:04:56.427089 7216 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI0312 18:04:56.427156 7216 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI0312 18:04:56.427246 7216 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 18:04:56.427294 7216 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 18:04:56.427311 7216 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 18:04:56.427313 7216 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 18:04:56.427343 7216 factory.go:656] Stopping watch factory\\\\nI0312 18:04:56.427360 7216 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 18:04:56.427397 7216 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 18:04:56.427659 7216 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 18:04:56.427747 7216 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 18:04:56.427790 7216 ovnkube.go:599] Stopped ovnkube\\\\nI0312 18:04:56.427823 7216 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0312 18:04:56.427914 7216 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:26Z\\\",\\\"message\\\":\\\"(nil)\\\\nI0312 18:05:26.683779 7541 services_controller.go:444] Built service openshift-dns/dns-default LB per-node configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"UDP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF0312 18:05:26.683797 7541 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.300407 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.315379 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.336000 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.355902 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.375889 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.391508 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.410285 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.426000 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.443021 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.457738 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.471598 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.488656 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.499249 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.502818 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.509490 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.523935 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:27 crc kubenswrapper[4926]: I0312 18:05:27.537519 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:27Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.259722 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/3.log" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.264593 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:05:28 crc kubenswrapper[4926]: E0312 18:05:28.264911 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.278751 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c31076f-c495-4763-9ea0-58712f07bb6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://405f59d0da6c9a3663ed746f08f9d5c2d94818971dbc0ce0373690c731b5afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:44Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 18:03:14.139597 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 18:03:14.141034 1 observer_polling.go:159] Starting file observer\\\\nI0312 18:03:14.142357 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 18:03:14.143332 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 18:03:43.393935 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0312 18:03:44.289189 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 18:03:44.289250 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:13Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0476f8d782654dd0fc67232862117105145e03f9a9495fb7b7015c28dac5c4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f33c0bfdb670b43186efb3e52df85915bd35749a127245356f71fe96994d85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8ff9f1fe3b91c7273624abf9e138c54d1d2228edc8e5ff370cdcc3b8df4a7d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.291134 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.300942 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"467073c1-7776-4dca-9a24-1beb51b5775f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a00032a5db7175b95edd80be3a15bca3e5bfee1c8bcc8bb2353ab3b620e12b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc024e5f387159c7533f0ba92814dda334990d955a31d8387cc943e31d4f6a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fc024e5f387159c7533f0ba92814dda334990d955a31d8387cc943e31d4f6a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.316667 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.334429 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.352179 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.369324 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.389214 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.404234 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.419449 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.436640 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.453315 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.474825 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:26Z\\\",\\\"message\\\":\\\"(nil)\\\\nI0312 18:05:26.683779 7541 services_controller.go:444] Built service openshift-dns/dns-default LB per-node configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"UDP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF0312 18:05:26.683797 7541 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:05:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.489251 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.489304 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:28 crc kubenswrapper[4926]: E0312 18:05:28.489410 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.489505 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:28 crc kubenswrapper[4926]: E0312 18:05:28.489617 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.489250 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:28 crc kubenswrapper[4926]: E0312 18:05:28.489706 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:28 crc kubenswrapper[4926]: E0312 18:05:28.489793 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.491667 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.502254 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.515754 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.531773 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:28 crc kubenswrapper[4926]: I0312 18:05:28.544187 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:28Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.328264 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.328567 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:34.328531653 +0000 UTC m=+234.697158036 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.329534 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.329732 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.329737 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.330096 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:06:34.33007683 +0000 UTC m=+234.698703193 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.329786 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.330509 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 18:06:34.330491592 +0000 UTC m=+234.699117955 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.430845 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.430909 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431052 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431094 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431114 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431140 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431168 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431238 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431187 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 18:06:34.431162905 +0000 UTC m=+234.799789268 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.431322 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 18:06:34.431298819 +0000 UTC m=+234.799925192 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.473241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.473311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.473334 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.473386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.473409 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:30Z","lastTransitionTime":"2026-03-12T18:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.489729 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.489839 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.489918 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.489923 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.489974 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.490127 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.490281 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.490402 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.493331 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.502956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.502991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.503007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.503021 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.503033 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:30Z","lastTransitionTime":"2026-03-12T18:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.509173 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.516506 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.521217 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.521290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.521305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.521320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.521329 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:30Z","lastTransitionTime":"2026-03-12T18:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.526786 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.535147 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.539158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.539197 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.539215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.539236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.539252 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:30Z","lastTransitionTime":"2026-03-12T18:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.541884 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.553699 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.557789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.557840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.557860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.557885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.557903 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:30Z","lastTransitionTime":"2026-03-12T18:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.563738 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:26Z\\\",\\\"message\\\":\\\"(nil)\\\\nI0312 18:05:26.683779 7541 services_controller.go:444] Built service openshift-dns/dns-default LB per-node configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"UDP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF0312 18:05:26.683797 7541 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:05:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.573139 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.573354 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.581499 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: E0312 18:05:30.590616 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.595885 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.614109 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.627093 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.644918 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.656793 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.668440 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.681583 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c31076f-c495-4763-9ea0-58712f07bb6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://405f59d0da6c9a3663ed746f08f9d5c2d94818971dbc0ce0373690c731b5afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:44Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 18:03:14.139597 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 18:03:14.141034 1 observer_polling.go:159] Starting file observer\\\\nI0312 18:03:14.142357 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 18:03:14.143332 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 18:03:43.393935 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0312 18:03:44.289189 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 18:03:44.289250 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:13Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0476f8d782654dd0fc67232862117105145e03f9a9495fb7b7015c28dac5c4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f33c0bfdb670b43186efb3e52df85915bd35749a127245356f71fe96994d85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8ff9f1fe3b91c7273624abf9e138c54d1d2228edc8e5ff370cdcc3b8df4a7d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.691289 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.703367 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"467073c1-7776-4dca-9a24-1beb51b5775f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a00032a5db7175b95edd80be3a15bca3e5bfee1c8bcc8bb2353ab3b620e12b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc024e5f387159c7533f0ba92814dda334990d955a31d8387cc943e31d4f6a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fc024e5f387159c7533f0ba92814dda334990d955a31d8387cc943e31d4f6a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.717615 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.736113 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.751509 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:30 crc kubenswrapper[4926]: I0312 18:05:30.765564 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:30Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:32 crc kubenswrapper[4926]: I0312 18:05:32.489360 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:32 crc kubenswrapper[4926]: I0312 18:05:32.489298 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:32 crc kubenswrapper[4926]: E0312 18:05:32.489890 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:32 crc kubenswrapper[4926]: I0312 18:05:32.489514 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:32 crc kubenswrapper[4926]: I0312 18:05:32.489478 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:32 crc kubenswrapper[4926]: E0312 18:05:32.489974 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:32 crc kubenswrapper[4926]: E0312 18:05:32.490154 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:32 crc kubenswrapper[4926]: E0312 18:05:32.490238 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:34 crc kubenswrapper[4926]: I0312 18:05:34.488956 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:34 crc kubenswrapper[4926]: I0312 18:05:34.489019 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:34 crc kubenswrapper[4926]: I0312 18:05:34.489007 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:34 crc kubenswrapper[4926]: I0312 18:05:34.488956 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:34 crc kubenswrapper[4926]: E0312 18:05:34.489209 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:34 crc kubenswrapper[4926]: E0312 18:05:34.489311 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:34 crc kubenswrapper[4926]: E0312 18:05:34.489558 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:34 crc kubenswrapper[4926]: E0312 18:05:34.489727 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:35 crc kubenswrapper[4926]: E0312 18:05:35.592356 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:36 crc kubenswrapper[4926]: I0312 18:05:36.489335 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:36 crc kubenswrapper[4926]: I0312 18:05:36.489413 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:36 crc kubenswrapper[4926]: E0312 18:05:36.489588 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:36 crc kubenswrapper[4926]: I0312 18:05:36.489627 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:36 crc kubenswrapper[4926]: I0312 18:05:36.489700 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:36 crc kubenswrapper[4926]: E0312 18:05:36.490002 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:36 crc kubenswrapper[4926]: E0312 18:05:36.490156 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:36 crc kubenswrapper[4926]: E0312 18:05:36.490288 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:36 crc kubenswrapper[4926]: I0312 18:05:36.508139 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 12 18:05:38 crc kubenswrapper[4926]: I0312 18:05:38.489763 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:38 crc kubenswrapper[4926]: I0312 18:05:38.489887 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:38 crc kubenswrapper[4926]: I0312 18:05:38.489967 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:38 crc kubenswrapper[4926]: I0312 18:05:38.490233 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:38 crc kubenswrapper[4926]: E0312 18:05:38.490246 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:38 crc kubenswrapper[4926]: E0312 18:05:38.490385 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:38 crc kubenswrapper[4926]: E0312 18:05:38.490497 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:38 crc kubenswrapper[4926]: E0312 18:05:38.490624 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.488915 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.489067 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.489130 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.489247 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.489328 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.489639 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.489748 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.489515 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.513535 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5c31076f-c495-4763-9ea0-58712f07bb6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://405f59d0da6c9a3663ed746f08f9d5c2d94818971dbc0ce0373690c731b5afae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ede7f52bcdc49e61c1fd76151da86db1e67d189c2fe147a60e207d4aa2dbfed8\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:44Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 18:03:14.139597 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 18:03:14.141034 1 observer_polling.go:159] Starting file observer\\\\nI0312 18:03:14.142357 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 18:03:14.143332 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 18:03:43.393935 1 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials\\\\nI0312 18:03:44.289189 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 18:03:44.289250 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:13Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:03:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0476f8d782654dd0fc67232862117105145e03f9a9495fb7b7015c28dac5c4dc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f33c0bfdb670b43186efb3e52df85915bd35749a127245356f71fe96994d85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8ff9f1fe3b91c7273624abf9e138c54d1d2228edc8e5ff370cdcc3b8df4a7d5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.524584 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4gmrt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dcfdbe34-faf7-4306-a2d8-6e95715f4f2a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aec75b239d7939d135db129ca6960660469ae8a708cfb841456f865c900fdfb5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fmvtd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:29Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4gmrt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.535995 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"467073c1-7776-4dca-9a24-1beb51b5775f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a00032a5db7175b95edd80be3a15bca3e5bfee1c8bcc8bb2353ab3b620e12b08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc024e5f387159c7533f0ba92814dda334990d955a31d8387cc943e31d4f6a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fc024e5f387159c7533f0ba92814dda334990d955a31d8387cc943e31d4f6a3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.553346 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.573552 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6084f41c213b6a4cdd05e3347e853818264c55f770f84083635e32dde284a489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.593653 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.597289 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.608875 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f7b34559-da2f-4796-8f3f-c56b2725c464\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a6ddc83150ed274605a4162acaba9d7326df9cb34f2197c177597af1cfc9d19\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8t62\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-hmdg8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.619551 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f369f51b-80f9-46fb-b43d-d6e057d3ebf3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:03:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16541dcf0547cc5940cc3d4d8b3007cd56ffd6fcbf3f7a042b641afeed488a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4d1c51341df9f02f1ffb0ac64cf07549f04ad8b349843cd11e756ca8d3be225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f881f53148e0d424118d6c39bcb9b736ec796d6bca00ca4a67714186d8c57c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90ff4bf24bc46ebc21c4b6f77467c73f094cca582e806b41b740ce17119226f2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.638355 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"751cfb31-c4cb-4d6a-8439-c2d4a64ccfe5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4da583451dead089b4282bf5a7ae151758817a7761131745a238968a4a1d9f75\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://587b28a1d4a204a36e5a2d1828a497d64dd9774cef86641c25dd276fe3185c12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59e6a444813d1bdcdb28520b832548d594b7170869334f17d0e1cb3c5c4c3338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://357166a9a18663c77154337ba5b95c2485ed2d008dab9caf790697c54e55da27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5b57f2b6ccf2e83dc3cdc2b924966715676cca996f717685dda64c8a1556525\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93765ed57dbcebda4b710bf4cafac472d4a67f6d60b259c2314ad5169301c1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93765ed57dbcebda4b710bf4cafac472d4a67f6d60b259c2314ad5169301c1e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7fb489c2b03ee964d954776552ca7b84509be8dfcfff33050c034d54a141bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7fb489c2b03ee964d954776552ca7b84509be8dfcfff33050c034d54a141bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://63c194a342c3a7d6ac13482ec064af4b05918784b371b742dd4d9e15a3c05cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63c194a342c3a7d6ac13482ec064af4b05918784b371b742dd4d9e15a3c05cb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.651997 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.667232 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13bb5191fe71e7796ec5d778fdafbd01d59eb57b8cf02c5e588a1a2714f371f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9db693cb7a535d2499f919ccbabcebc5b97c340fe0dae0a02dc03121b24c9ee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.687189 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc33af41-5aa0-4254-ac75-69433d5f4ce9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:26Z\\\",\\\"message\\\":\\\"(nil)\\\\nI0312 18:05:26.683779 7541 services_controller.go:444] Built service openshift-dns/dns-default LB per-node configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"UDP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:53, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.4.10\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9154, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF0312 18:05:26.683797 7541 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:05:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4t5dd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zlfmg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.701839 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srh42" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1d37aa11-8fa5-4eb3-8edd-6f71523623b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eda521cfe2ce4a398e537a7f447b65616af12bce78d6fad3e4aafa2a34195eb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b10730077cd07faf61ad5167a5ee7ef1630bd4b48dd478080604f8f3bd511818\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73c4731a9a0bb1d7c53c7ffcdb7e6ed42840c6e2b4247cbff3aad6f2e7b39101\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://099d0eabc51c98a9d49bdf0ea3a18627e138b6a83d5769fd8d62467de933324e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bae207872238453e4fbb97dbed373dd96b22348248989c66074062988d0cda2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c69f3072bac6f67905ba48b8f894b0d682e66c58ea7749c96507be310bee09a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1fe92341e742b5099f178691183a9e139eb4f129c53ff3a7ff9c4a8731599ef\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:04:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbt59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srh42\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.712114 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"211eeae6-9b41-484b-bd13-99c1c28cdf96\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n6bnm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-n7pd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.723791 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aeb621bb-05ee-456b-b869-1cdd14184ad1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:02:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T18:03:48Z\\\",\\\"message\\\":\\\"W0312 18:03:47.732255 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0312 18:03:47.732907 1 crypto.go:601] Generating new CA for check-endpoints-signer@1773338627 cert, and key in /tmp/serving-cert-406653352/serving-signer.crt, /tmp/serving-cert-406653352/serving-signer.key\\\\nI0312 18:03:48.334309 1 observer_polling.go:159] Starting file observer\\\\nW0312 18:03:48.343915 1 builder.go:272] unable to get owner reference (falling back to namespace): Unauthorized\\\\nI0312 18:03:48.344098 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 18:03:48.345204 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-406653352/tls.crt::/tmp/serving-cert-406653352/tls.key\\\\\\\"\\\\nF0312 18:03:48.873093 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:03:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:02:43Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T18:02:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T18:02:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:02:40Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.735840 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:30Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8dcba7a43670fd5cf236b71365bc171373a7d2eb9bc24c01f5e8e49260a359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.749770 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xwqvl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5a53ef4-c701-457f-9cf2-85819bf04d1a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T18:05:14Z\\\",\\\"message\\\":\\\"2026-03-12T18:04:28+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d\\\\n2026-03-12T18:04:28+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_38286ed1-c497-4a74-9174-f8362ccb899d to /host/opt/cni/bin/\\\\n2026-03-12T18:04:29Z [verbose] multus-daemon started\\\\n2026-03-12T18:04:29Z [verbose] Readiness Indicator file check\\\\n2026-03-12T18:05:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:05:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bt6hw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xwqvl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.760675 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-f9vxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"594c806d-dd79-41ce-8e3a-a33d42bf0f7e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://690b46808ceaeee6e4769dad24c7bdb281b5530a27d5079e9ddf1e55f914171e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x7bdm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:26Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-f9vxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.769062 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.769200 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.769272 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs podName:211eeae6-9b41-484b-bd13-99c1c28cdf96 nodeName:}" failed. No retries permitted until 2026-03-12 18:06:44.769254151 +0000 UTC m=+245.137880574 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs") pod "network-metrics-daemon-n7pd7" (UID: "211eeae6-9b41-484b-bd13-99c1c28cdf96") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.771520 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12de8a94-72e6-4d72-8e39-42f3ef9d1125\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T18:04:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f1c4af6ac2ce0a767b4d67627bcab2f7617b3c7845fbcb3136f3cf8931dc186\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7342dbad64c6f7552247607ae0c5da9b7490241867cb0a212d424e2b581910a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T18:04:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vwn2v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T18:04:35Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fq9dc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.970753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.970818 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.970834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.970850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.970862 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:40Z","lastTransitionTime":"2026-03-12T18:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:40 crc kubenswrapper[4926]: E0312 18:05:40.989570 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:40Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.993627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.993672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.993682 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.993699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:40 crc kubenswrapper[4926]: I0312 18:05:40.993713 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:40Z","lastTransitionTime":"2026-03-12T18:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:41 crc kubenswrapper[4926]: E0312 18:05:41.007604 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.011588 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.011630 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.011641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.011658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.011674 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:41Z","lastTransitionTime":"2026-03-12T18:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:41 crc kubenswrapper[4926]: E0312 18:05:41.025180 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.029478 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.029529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.029544 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.029564 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.029577 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:41Z","lastTransitionTime":"2026-03-12T18:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:41 crc kubenswrapper[4926]: E0312 18:05:41.046471 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.050648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.050698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.050707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.050726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:41 crc kubenswrapper[4926]: I0312 18:05:41.050736 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:41Z","lastTransitionTime":"2026-03-12T18:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:41 crc kubenswrapper[4926]: E0312 18:05:41.067109 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T18:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2090c8b2-af81-407e-bc9b-78510eed61ed\\\",\\\"systemUUID\\\":\\\"9f4a0cfb-e2ee-40d1-a613-eac4618fc62c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T18:05:41Z is after 2025-08-24T17:21:41Z" Mar 12 18:05:41 crc kubenswrapper[4926]: E0312 18:05:41.067235 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 18:05:42 crc kubenswrapper[4926]: I0312 18:05:42.488950 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:42 crc kubenswrapper[4926]: I0312 18:05:42.488994 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:42 crc kubenswrapper[4926]: E0312 18:05:42.489149 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:42 crc kubenswrapper[4926]: I0312 18:05:42.489183 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:42 crc kubenswrapper[4926]: E0312 18:05:42.489356 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:42 crc kubenswrapper[4926]: I0312 18:05:42.489489 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:42 crc kubenswrapper[4926]: E0312 18:05:42.489506 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:42 crc kubenswrapper[4926]: E0312 18:05:42.489640 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:43 crc kubenswrapper[4926]: I0312 18:05:43.490888 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:05:43 crc kubenswrapper[4926]: E0312 18:05:43.491334 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:05:44 crc kubenswrapper[4926]: I0312 18:05:44.489199 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:44 crc kubenswrapper[4926]: I0312 18:05:44.489278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:44 crc kubenswrapper[4926]: I0312 18:05:44.489232 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:44 crc kubenswrapper[4926]: E0312 18:05:44.489376 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:44 crc kubenswrapper[4926]: E0312 18:05:44.489551 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:44 crc kubenswrapper[4926]: E0312 18:05:44.489624 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:44 crc kubenswrapper[4926]: I0312 18:05:44.489755 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:44 crc kubenswrapper[4926]: E0312 18:05:44.490079 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:45 crc kubenswrapper[4926]: E0312 18:05:45.595032 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:46 crc kubenswrapper[4926]: I0312 18:05:46.489951 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:46 crc kubenswrapper[4926]: I0312 18:05:46.490030 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:46 crc kubenswrapper[4926]: E0312 18:05:46.490122 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:46 crc kubenswrapper[4926]: I0312 18:05:46.490225 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:46 crc kubenswrapper[4926]: E0312 18:05:46.490269 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:46 crc kubenswrapper[4926]: I0312 18:05:46.490414 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:46 crc kubenswrapper[4926]: E0312 18:05:46.490517 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:46 crc kubenswrapper[4926]: E0312 18:05:46.490668 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:48 crc kubenswrapper[4926]: I0312 18:05:48.489171 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:48 crc kubenswrapper[4926]: I0312 18:05:48.489241 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:48 crc kubenswrapper[4926]: I0312 18:05:48.489257 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:48 crc kubenswrapper[4926]: E0312 18:05:48.489338 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:48 crc kubenswrapper[4926]: I0312 18:05:48.489491 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:48 crc kubenswrapper[4926]: E0312 18:05:48.489494 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:48 crc kubenswrapper[4926]: E0312 18:05:48.489602 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:48 crc kubenswrapper[4926]: E0312 18:05:48.489775 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.489336 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.489336 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:50 crc kubenswrapper[4926]: E0312 18:05:50.489472 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.489501 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.489423 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:50 crc kubenswrapper[4926]: E0312 18:05:50.489609 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:50 crc kubenswrapper[4926]: E0312 18:05:50.489693 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:50 crc kubenswrapper[4926]: E0312 18:05:50.489742 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.520931 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=23.520898785 podStartE2EDuration="23.520898785s" podCreationTimestamp="2026-03-12 18:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.520670229 +0000 UTC m=+190.889296562" watchObservedRunningTime="2026-03-12 18:05:50.520898785 +0000 UTC m=+190.889525158" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.552064 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-4gmrt" podStartSLOduration=119.552031218 podStartE2EDuration="1m59.552031218s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.536214449 +0000 UTC m=+190.904840782" watchObservedRunningTime="2026-03-12 18:05:50.552031218 +0000 UTC m=+190.920657551" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.574347 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podStartSLOduration=119.574304818 podStartE2EDuration="1m59.574304818s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.562846428 +0000 UTC m=+190.931472771" watchObservedRunningTime="2026-03-12 18:05:50.574304818 +0000 UTC m=+190.942931151" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.575013 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=23.575005098 podStartE2EDuration="23.575005098s" podCreationTimestamp="2026-03-12 18:05:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.574247416 +0000 UTC m=+190.942873749" watchObservedRunningTime="2026-03-12 18:05:50.575005098 +0000 UTC m=+190.943631431" Mar 12 18:05:50 crc kubenswrapper[4926]: E0312 18:05:50.596533 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.680146 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-srh42" podStartSLOduration=119.680130502 podStartE2EDuration="1m59.680130502s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.678196625 +0000 UTC m=+191.046822958" watchObservedRunningTime="2026-03-12 18:05:50.680130502 +0000 UTC m=+191.048756835" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.710929 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=72.710906044 podStartE2EDuration="1m12.710906044s" podCreationTimestamp="2026-03-12 18:04:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.710606275 +0000 UTC m=+191.079232608" watchObservedRunningTime="2026-03-12 18:05:50.710906044 +0000 UTC m=+191.079532377" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.734521 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=14.734499652 podStartE2EDuration="14.734499652s" podCreationTimestamp="2026-03-12 18:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.734141443 +0000 UTC m=+191.102767776" watchObservedRunningTime="2026-03-12 18:05:50.734499652 +0000 UTC m=+191.103125995" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.769687 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-f9vxh" podStartSLOduration=119.769663545 podStartE2EDuration="1m59.769663545s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.757259017 +0000 UTC m=+191.125885350" watchObservedRunningTime="2026-03-12 18:05:50.769663545 +0000 UTC m=+191.138289918" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.793680 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fq9dc" podStartSLOduration=119.793658895 podStartE2EDuration="1m59.793658895s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.769993915 +0000 UTC m=+191.138620248" watchObservedRunningTime="2026-03-12 18:05:50.793658895 +0000 UTC m=+191.162285248" Mar 12 18:05:50 crc kubenswrapper[4926]: I0312 18:05:50.794010 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.794004235 podStartE2EDuration="1m21.794004235s" podCreationTimestamp="2026-03-12 18:04:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.792706357 +0000 UTC m=+191.161332700" watchObservedRunningTime="2026-03-12 18:05:50.794004235 +0000 UTC m=+191.162630578" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.216233 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.216583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.216727 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.216864 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.216992 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T18:05:51Z","lastTransitionTime":"2026-03-12T18:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.270724 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xwqvl" podStartSLOduration=120.270694997 podStartE2EDuration="2m0.270694997s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:50.876945273 +0000 UTC m=+191.245571596" watchObservedRunningTime="2026-03-12 18:05:51.270694997 +0000 UTC m=+191.639321370" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.272275 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh"] Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.273085 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.276635 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.276922 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.277107 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.277333 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.285051 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.285110 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.285265 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.285295 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.285367 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.386230 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.387062 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.387128 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.387151 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.387174 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.387416 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.387520 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.388137 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-service-ca\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.396537 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.407754 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a27d1b4e-7996-4cbe-a080-c3fbc9f865ef-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-s7lzh\" (UID: \"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.543928 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.551523 4926 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 12 18:05:51 crc kubenswrapper[4926]: I0312 18:05:51.585314 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" Mar 12 18:05:52 crc kubenswrapper[4926]: I0312 18:05:52.349889 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" event={"ID":"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef","Type":"ContainerStarted","Data":"07414689799ff8b48532f3abea063045d7db46c7a9af2329b1e92826e7d27a8a"} Mar 12 18:05:52 crc kubenswrapper[4926]: I0312 18:05:52.350200 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" event={"ID":"a27d1b4e-7996-4cbe-a080-c3fbc9f865ef","Type":"ContainerStarted","Data":"7a9c9a091fc5f64dc99b67261f0cdb0d61c11b9d3fcb0cc7436de459525e3eda"} Mar 12 18:05:52 crc kubenswrapper[4926]: I0312 18:05:52.367735 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-s7lzh" podStartSLOduration=121.367714334 podStartE2EDuration="2m1.367714334s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:05:52.367095126 +0000 UTC m=+192.735721459" watchObservedRunningTime="2026-03-12 18:05:52.367714334 +0000 UTC m=+192.736340667" Mar 12 18:05:52 crc kubenswrapper[4926]: I0312 18:05:52.489616 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:52 crc kubenswrapper[4926]: I0312 18:05:52.489657 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:52 crc kubenswrapper[4926]: I0312 18:05:52.489696 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:52 crc kubenswrapper[4926]: I0312 18:05:52.489712 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:52 crc kubenswrapper[4926]: E0312 18:05:52.490180 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:52 crc kubenswrapper[4926]: E0312 18:05:52.490320 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:52 crc kubenswrapper[4926]: E0312 18:05:52.490356 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:52 crc kubenswrapper[4926]: E0312 18:05:52.490481 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:54 crc kubenswrapper[4926]: I0312 18:05:54.489020 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:54 crc kubenswrapper[4926]: E0312 18:05:54.489238 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:54 crc kubenswrapper[4926]: I0312 18:05:54.489279 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:54 crc kubenswrapper[4926]: I0312 18:05:54.489292 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:54 crc kubenswrapper[4926]: E0312 18:05:54.489646 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:54 crc kubenswrapper[4926]: I0312 18:05:54.489953 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:54 crc kubenswrapper[4926]: E0312 18:05:54.490085 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:54 crc kubenswrapper[4926]: E0312 18:05:54.490995 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:54 crc kubenswrapper[4926]: I0312 18:05:54.491431 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:05:54 crc kubenswrapper[4926]: E0312 18:05:54.491687 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zlfmg_openshift-ovn-kubernetes(bc33af41-5aa0-4254-ac75-69433d5f4ce9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" Mar 12 18:05:55 crc kubenswrapper[4926]: E0312 18:05:55.598563 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:05:56 crc kubenswrapper[4926]: I0312 18:05:56.489992 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:56 crc kubenswrapper[4926]: I0312 18:05:56.490060 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:56 crc kubenswrapper[4926]: I0312 18:05:56.489993 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:56 crc kubenswrapper[4926]: E0312 18:05:56.490138 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:05:56 crc kubenswrapper[4926]: I0312 18:05:56.490207 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:56 crc kubenswrapper[4926]: E0312 18:05:56.490271 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:56 crc kubenswrapper[4926]: E0312 18:05:56.490369 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:56 crc kubenswrapper[4926]: E0312 18:05:56.490423 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:58 crc kubenswrapper[4926]: I0312 18:05:58.489088 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:05:58 crc kubenswrapper[4926]: E0312 18:05:58.490180 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:05:58 crc kubenswrapper[4926]: I0312 18:05:58.489281 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:05:58 crc kubenswrapper[4926]: E0312 18:05:58.490380 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:05:58 crc kubenswrapper[4926]: I0312 18:05:58.489247 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:05:58 crc kubenswrapper[4926]: I0312 18:05:58.489316 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:05:58 crc kubenswrapper[4926]: E0312 18:05:58.490618 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:05:58 crc kubenswrapper[4926]: E0312 18:05:58.490842 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.378230 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/1.log" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.379108 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/0.log" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.379165 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5a53ef4-c701-457f-9cf2-85819bf04d1a" containerID="bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4" exitCode=1 Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.379198 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerDied","Data":"bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4"} Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.379231 4926 scope.go:117] "RemoveContainer" containerID="54fa97a4cd767400fe757a9f362d3aa29c51699ebd5671c32a04d7fbad9d6c6c" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.379552 4926 scope.go:117] "RemoveContainer" containerID="bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4" Mar 12 18:06:00 crc kubenswrapper[4926]: E0312 18:06:00.379706 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xwqvl_openshift-multus(d5a53ef4-c701-457f-9cf2-85819bf04d1a)\"" pod="openshift-multus/multus-xwqvl" podUID="d5a53ef4-c701-457f-9cf2-85819bf04d1a" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.489064 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.489106 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.489526 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:00 crc kubenswrapper[4926]: E0312 18:06:00.490465 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:00 crc kubenswrapper[4926]: I0312 18:06:00.490531 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:00 crc kubenswrapper[4926]: E0312 18:06:00.490691 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:00 crc kubenswrapper[4926]: E0312 18:06:00.490784 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:00 crc kubenswrapper[4926]: E0312 18:06:00.490997 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:00 crc kubenswrapper[4926]: E0312 18:06:00.600043 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:06:01 crc kubenswrapper[4926]: I0312 18:06:01.383990 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/1.log" Mar 12 18:06:02 crc kubenswrapper[4926]: I0312 18:06:02.489387 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:02 crc kubenswrapper[4926]: E0312 18:06:02.489596 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:02 crc kubenswrapper[4926]: I0312 18:06:02.489863 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:02 crc kubenswrapper[4926]: E0312 18:06:02.489978 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:02 crc kubenswrapper[4926]: I0312 18:06:02.490193 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:02 crc kubenswrapper[4926]: E0312 18:06:02.490299 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:02 crc kubenswrapper[4926]: I0312 18:06:02.490605 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:02 crc kubenswrapper[4926]: E0312 18:06:02.490747 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:04 crc kubenswrapper[4926]: I0312 18:06:04.489282 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:04 crc kubenswrapper[4926]: I0312 18:06:04.489364 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:04 crc kubenswrapper[4926]: I0312 18:06:04.489282 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:04 crc kubenswrapper[4926]: E0312 18:06:04.489553 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:04 crc kubenswrapper[4926]: E0312 18:06:04.489681 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:04 crc kubenswrapper[4926]: E0312 18:06:04.489925 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:04 crc kubenswrapper[4926]: I0312 18:06:04.490678 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:04 crc kubenswrapper[4926]: E0312 18:06:04.490792 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:05 crc kubenswrapper[4926]: E0312 18:06:05.601585 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:06:06 crc kubenswrapper[4926]: I0312 18:06:06.489891 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:06 crc kubenswrapper[4926]: I0312 18:06:06.489919 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:06 crc kubenswrapper[4926]: I0312 18:06:06.489974 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:06 crc kubenswrapper[4926]: E0312 18:06:06.490728 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:06 crc kubenswrapper[4926]: E0312 18:06:06.490825 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:06 crc kubenswrapper[4926]: I0312 18:06:06.489993 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:06 crc kubenswrapper[4926]: E0312 18:06:06.490982 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:06 crc kubenswrapper[4926]: E0312 18:06:06.491284 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:07 crc kubenswrapper[4926]: I0312 18:06:07.490137 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.407279 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/3.log" Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.409582 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerStarted","Data":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.410009 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.440594 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n7pd7"] Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.440691 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:08 crc kubenswrapper[4926]: E0312 18:06:08.440762 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.489639 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.489751 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:08 crc kubenswrapper[4926]: I0312 18:06:08.489781 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:08 crc kubenswrapper[4926]: E0312 18:06:08.490095 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:08 crc kubenswrapper[4926]: E0312 18:06:08.489940 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:08 crc kubenswrapper[4926]: E0312 18:06:08.490200 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:10 crc kubenswrapper[4926]: I0312 18:06:10.488927 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:10 crc kubenswrapper[4926]: I0312 18:06:10.490984 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:10 crc kubenswrapper[4926]: I0312 18:06:10.491011 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:10 crc kubenswrapper[4926]: I0312 18:06:10.491050 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:10 crc kubenswrapper[4926]: E0312 18:06:10.491131 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:10 crc kubenswrapper[4926]: E0312 18:06:10.491230 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:10 crc kubenswrapper[4926]: E0312 18:06:10.491312 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:10 crc kubenswrapper[4926]: E0312 18:06:10.491815 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:10 crc kubenswrapper[4926]: E0312 18:06:10.603029 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:06:12 crc kubenswrapper[4926]: I0312 18:06:12.489838 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:12 crc kubenswrapper[4926]: I0312 18:06:12.489929 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:12 crc kubenswrapper[4926]: E0312 18:06:12.490054 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:12 crc kubenswrapper[4926]: I0312 18:06:12.490083 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:12 crc kubenswrapper[4926]: E0312 18:06:12.490226 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:12 crc kubenswrapper[4926]: I0312 18:06:12.490294 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:12 crc kubenswrapper[4926]: E0312 18:06:12.490536 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:12 crc kubenswrapper[4926]: E0312 18:06:12.491864 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:12 crc kubenswrapper[4926]: I0312 18:06:12.492480 4926 scope.go:117] "RemoveContainer" containerID="bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4" Mar 12 18:06:12 crc kubenswrapper[4926]: I0312 18:06:12.515220 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podStartSLOduration=141.515197578 podStartE2EDuration="2m21.515197578s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:08.44646175 +0000 UTC m=+208.815088103" watchObservedRunningTime="2026-03-12 18:06:12.515197578 +0000 UTC m=+212.883823951" Mar 12 18:06:13 crc kubenswrapper[4926]: I0312 18:06:13.431768 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/1.log" Mar 12 18:06:13 crc kubenswrapper[4926]: I0312 18:06:13.432130 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerStarted","Data":"2c510c831017eeb7aed88601040d4790cf6d5bebce7b09f246c92ea9b81e2481"} Mar 12 18:06:14 crc kubenswrapper[4926]: I0312 18:06:14.489906 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:14 crc kubenswrapper[4926]: I0312 18:06:14.489949 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:14 crc kubenswrapper[4926]: I0312 18:06:14.490005 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:14 crc kubenswrapper[4926]: E0312 18:06:14.490038 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 18:06:14 crc kubenswrapper[4926]: I0312 18:06:14.490059 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:14 crc kubenswrapper[4926]: E0312 18:06:14.490179 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-n7pd7" podUID="211eeae6-9b41-484b-bd13-99c1c28cdf96" Mar 12 18:06:14 crc kubenswrapper[4926]: E0312 18:06:14.490220 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 18:06:14 crc kubenswrapper[4926]: E0312 18:06:14.490272 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.489068 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.490077 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.490309 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.490376 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.493045 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.493312 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.494754 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.494846 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.495833 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 12 18:06:16 crc kubenswrapper[4926]: I0312 18:06:16.496077 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.232606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.337764 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.338379 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.338580 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.338726 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qs58t"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.339047 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.339197 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vb9qx"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.339474 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.339807 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.340273 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bfn44"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.340732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.342175 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-z95pp"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.342956 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.349560 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350101 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350269 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350281 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350484 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350621 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350658 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350845 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351031 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351078 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.350629 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351197 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351268 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351399 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351713 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351817 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.351851 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.352029 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.353788 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.354331 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.354879 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.354893 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.355220 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.360902 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.361378 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.361535 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.361543 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.361711 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.361885 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.363312 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.363536 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.363684 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.363826 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.363934 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.364030 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.370922 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.371129 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.371622 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.371791 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.371796 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.372487 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.374980 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s9tvm"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.375825 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377069 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377156 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377322 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377399 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377621 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377715 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377909 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.377954 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.378096 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.378279 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.378511 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.378699 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.378803 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.378904 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.379007 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.380726 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7rmsn"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.383583 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.395962 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.401604 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.402662 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.403254 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.407930 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.410047 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.410634 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.411349 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2wsbc"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.412227 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.412644 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.412870 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-st67x"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.412889 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413165 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413272 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj7xf\" (UniqueName: \"kubernetes.io/projected/7e94e6e0-b16d-462f-b791-ba20acdcb809-kube-api-access-fj7xf\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413301 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c30a7ad-f92b-445a-9201-fe55f247cf41-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413332 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-config\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413348 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snkcw\" (UniqueName: \"kubernetes.io/projected/48a0fa25-6b2d-4668-b8e5-824912077f19-kube-api-access-snkcw\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413386 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5e304-df7c-434b-8b17-f520e9bb7d52-config\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413402 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/49e5e304-df7c-434b-8b17-f520e9bb7d52-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413417 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-serving-cert\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413433 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413469 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-serving-cert\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413485 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413509 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28x4l\" (UniqueName: \"kubernetes.io/projected/4230f869-9456-44a1-87b3-342fc8c18ed7-kube-api-access-28x4l\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413522 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-audit\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413537 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-encryption-config\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413554 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w622m\" (UniqueName: \"kubernetes.io/projected/4c2f00a8-c3ce-4957-9093-d8c2cce49992-kube-api-access-w622m\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413569 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e94e6e0-b16d-462f-b791-ba20acdcb809-audit-dir\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-config\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413606 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413623 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48a0fa25-6b2d-4668-b8e5-824912077f19-audit-dir\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413691 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-config\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413707 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t46vq\" (UniqueName: \"kubernetes.io/projected/270031fa-3d83-4edf-bb5d-19ce9e1a693d-kube-api-access-t46vq\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413768 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bh26\" (UniqueName: \"kubernetes.io/projected/49e5e304-df7c-434b-8b17-f520e9bb7d52-kube-api-access-6bh26\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413787 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-serving-cert\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413813 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-policies\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413849 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-serving-cert\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413870 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-dir\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413884 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413918 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-service-ca\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413937 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4230f869-9456-44a1-87b3-342fc8c18ed7-serving-cert\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413952 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55b6dd11-c219-4b24-90eb-dbc096a67835-config\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413972 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-oauth-config\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413989 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-config\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.413993 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-st67x" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414011 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414027 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414045 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-client\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414074 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414105 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-trusted-ca-bundle\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414123 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2f00a8-c3ce-4957-9093-d8c2cce49992-serving-cert\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414138 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-etcd-client\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414190 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/49e5e304-df7c-434b-8b17-f520e9bb7d52-images\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414211 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414232 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqbzk\" (UniqueName: \"kubernetes.io/projected/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-kube-api-access-hqbzk\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414252 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-config\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414269 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88hg4\" (UniqueName: \"kubernetes.io/projected/eedd886d-5443-47e1-afbf-5aff90067f3b-kube-api-access-88hg4\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414284 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9zg78"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414293 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/48a0fa25-6b2d-4668-b8e5-824912077f19-node-pullsecrets\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414313 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62f9n\" (UniqueName: \"kubernetes.io/projected/5db62dca-ba86-4ca4-861e-003d09e5ac0f-kube-api-access-62f9n\") pod \"cluster-samples-operator-665b6dd947-hwwdz\" (UID: \"5db62dca-ba86-4ca4-861e-003d09e5ac0f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414342 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414359 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414408 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5db62dca-ba86-4ca4-861e-003d09e5ac0f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hwwdz\" (UID: \"5db62dca-ba86-4ca4-861e-003d09e5ac0f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414429 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414473 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-image-import-ca\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414492 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-service-ca\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414521 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-etcd-client\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414539 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414556 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbgdh\" (UniqueName: \"kubernetes.io/projected/55b6dd11-c219-4b24-90eb-dbc096a67835-kube-api-access-jbgdh\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414578 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414601 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c30a7ad-f92b-445a-9201-fe55f247cf41-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414620 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-oauth-serving-cert\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414651 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-audit-policies\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414669 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7dxn\" (UniqueName: \"kubernetes.io/projected/8c30a7ad-f92b-445a-9201-fe55f247cf41-kube-api-access-c7dxn\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414727 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414774 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/55b6dd11-c219-4b24-90eb-dbc096a67835-machine-approver-tls\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414791 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414812 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-etcd-serving-ca\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414851 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-trusted-ca-bundle\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414886 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-ca\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414888 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414928 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/55b6dd11-c219-4b24-90eb-dbc096a67835-auth-proxy-config\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.414981 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-client-ca\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.415015 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5ctp\" (UniqueName: \"kubernetes.io/projected/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-kube-api-access-n5ctp\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.415035 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-encryption-config\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.415414 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h95sl"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.415965 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.417164 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q49l5"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.420040 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.420291 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.420510 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.420688 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.422669 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.422848 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.423147 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.424209 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.424721 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.424755 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.424918 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.424972 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.425092 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.425123 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.425225 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.425303 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.425425 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.425869 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g6fzt"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.426308 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.426826 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.427163 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bsvxw"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.427646 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.427875 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.427969 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.428111 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.428342 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.428492 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.428621 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.428879 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.428907 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.428949 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.429072 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.429432 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.429077 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.431425 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.432056 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.432182 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.432476 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433118 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433350 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433414 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433561 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433696 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433727 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433851 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.433981 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.434065 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.434186 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.434219 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.434417 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.461923 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-wndvq"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.462677 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.466724 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.468030 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.468518 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.469246 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.469316 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.469246 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.469695 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.469713 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.469923 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.470431 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.470804 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.471713 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.486226 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.486309 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.486881 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.487581 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.488079 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.488927 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c68kr"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.505811 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.508313 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.508911 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.509579 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.511266 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.511505 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.511786 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.514580 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.515620 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.515724 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517062 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-serving-cert\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bh26\" (UniqueName: \"kubernetes.io/projected/49e5e304-df7c-434b-8b17-f520e9bb7d52-kube-api-access-6bh26\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517137 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfrhq\" (UniqueName: \"kubernetes.io/projected/1dc82997-2782-4e9e-a293-956fcb96acde-kube-api-access-xfrhq\") pod \"dns-operator-744455d44c-9zg78\" (UID: \"1dc82997-2782-4e9e-a293-956fcb96acde\") " pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517156 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517176 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-policies\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517192 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-service-ca\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517209 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-serving-cert\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517226 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-dir\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517241 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517259 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1dc82997-2782-4e9e-a293-956fcb96acde-metrics-tls\") pod \"dns-operator-744455d44c-9zg78\" (UID: \"1dc82997-2782-4e9e-a293-956fcb96acde\") " pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517279 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4230f869-9456-44a1-87b3-342fc8c18ed7-serving-cert\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517296 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55b6dd11-c219-4b24-90eb-dbc096a67835-config\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517312 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-oauth-config\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517329 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ee39302-0316-4481-871e-538ffd31a507-trusted-ca\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517346 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517363 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517379 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-config\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517396 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517457 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2f00a8-c3ce-4957-9093-d8c2cce49992-serving-cert\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517478 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-client\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517493 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517509 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-trusted-ca-bundle\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517525 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqbzk\" (UniqueName: \"kubernetes.io/projected/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-kube-api-access-hqbzk\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-etcd-client\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517558 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct49b\" (UniqueName: \"kubernetes.io/projected/e379fe1d-7780-4f17-8df8-f74f3dddbc23-kube-api-access-ct49b\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517575 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/49e5e304-df7c-434b-8b17-f520e9bb7d52-images\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517591 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517607 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-config\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517625 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-stats-auth\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517644 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-config\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517660 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88hg4\" (UniqueName: \"kubernetes.io/projected/eedd886d-5443-47e1-afbf-5aff90067f3b-kube-api-access-88hg4\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517679 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62f9n\" (UniqueName: \"kubernetes.io/projected/5db62dca-ba86-4ca4-861e-003d09e5ac0f-kube-api-access-62f9n\") pod \"cluster-samples-operator-665b6dd947-hwwdz\" (UID: \"5db62dca-ba86-4ca4-861e-003d09e5ac0f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517695 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/48a0fa25-6b2d-4668-b8e5-824912077f19-node-pullsecrets\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517713 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517730 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517751 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517771 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517788 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ee39302-0316-4481-871e-538ffd31a507-serving-cert\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517817 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5db62dca-ba86-4ca4-861e-003d09e5ac0f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hwwdz\" (UID: \"5db62dca-ba86-4ca4-861e-003d09e5ac0f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517832 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-service-ca\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517849 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-image-import-ca\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517871 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-etcd-client\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517896 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517917 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c30a7ad-f92b-445a-9201-fe55f247cf41-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517932 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbgdh\" (UniqueName: \"kubernetes.io/projected/55b6dd11-c219-4b24-90eb-dbc096a67835-kube-api-access-jbgdh\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517948 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517965 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7dxn\" (UniqueName: \"kubernetes.io/projected/8c30a7ad-f92b-445a-9201-fe55f247cf41-kube-api-access-c7dxn\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517979 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-oauth-serving-cert\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.517996 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjftj\" (UniqueName: \"kubernetes.io/projected/7ee39302-0316-4481-871e-538ffd31a507-kube-api-access-tjftj\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518011 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-default-certificate\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518027 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-audit-policies\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518070 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e379fe1d-7780-4f17-8df8-f74f3dddbc23-service-ca-bundle\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518094 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/55b6dd11-c219-4b24-90eb-dbc096a67835-machine-approver-tls\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518153 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-ca\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518168 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-etcd-serving-ca\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518198 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-trusted-ca-bundle\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518215 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/55b6dd11-c219-4b24-90eb-dbc096a67835-auth-proxy-config\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518237 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-client-ca\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518257 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-encryption-config\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518276 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee39302-0316-4481-871e-538ffd31a507-config\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518297 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5ctp\" (UniqueName: \"kubernetes.io/projected/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-kube-api-access-n5ctp\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518321 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518342 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj7xf\" (UniqueName: \"kubernetes.io/projected/7e94e6e0-b16d-462f-b791-ba20acdcb809-kube-api-access-fj7xf\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518363 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c30a7ad-f92b-445a-9201-fe55f247cf41-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518381 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-config\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518398 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snkcw\" (UniqueName: \"kubernetes.io/projected/48a0fa25-6b2d-4668-b8e5-824912077f19-kube-api-access-snkcw\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518415 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5e304-df7c-434b-8b17-f520e9bb7d52-config\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518431 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/49e5e304-df7c-434b-8b17-f520e9bb7d52-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518465 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-serving-cert\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518486 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518503 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-serving-cert\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518521 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518538 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518557 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28x4l\" (UniqueName: \"kubernetes.io/projected/4230f869-9456-44a1-87b3-342fc8c18ed7-kube-api-access-28x4l\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518572 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-audit\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518587 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-encryption-config\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518603 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w622m\" (UniqueName: \"kubernetes.io/projected/4c2f00a8-c3ce-4957-9093-d8c2cce49992-kube-api-access-w622m\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e94e6e0-b16d-462f-b791-ba20acdcb809-audit-dir\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518636 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-config\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518652 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48a0fa25-6b2d-4668-b8e5-824912077f19-audit-dir\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518676 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-config\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518691 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-metrics-certs\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.518708 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t46vq\" (UniqueName: \"kubernetes.io/projected/270031fa-3d83-4edf-bb5d-19ce9e1a693d-kube-api-access-t46vq\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.519516 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/49e5e304-df7c-434b-8b17-f520e9bb7d52-images\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.519927 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-dir\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.521739 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.532176 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-ca\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.533198 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.533653 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-config\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.533863 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/48a0fa25-6b2d-4668-b8e5-824912077f19-node-pullsecrets\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.534860 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-etcd-serving-ca\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.535874 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-trusted-ca-bundle\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.536406 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-audit\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.536770 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/55b6dd11-c219-4b24-90eb-dbc096a67835-auth-proxy-config\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.547862 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.550004 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-policies\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.550083 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-trusted-ca-bundle\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.550216 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-client-ca\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.550482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-service-ca\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.550657 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-oauth-config\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.551172 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/49e5e304-df7c-434b-8b17-f520e9bb7d52-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.551598 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-config\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.551810 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-config\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.551961 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7e94e6e0-b16d-462f-b791-ba20acdcb809-audit-dir\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.552483 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-service-ca\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.552642 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.552879 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.559857 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.560015 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.560868 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.561062 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.561096 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.561794 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c30a7ad-f92b-445a-9201-fe55f247cf41-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.562193 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-config\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.562263 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7e94e6e0-b16d-462f-b791-ba20acdcb809-audit-policies\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.562836 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e5e304-df7c-434b-8b17-f520e9bb7d52-config\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.563223 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-etcd-client\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.564814 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-oauth-serving-cert\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.565313 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-config\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.565360 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48a0fa25-6b2d-4668-b8e5-824912077f19-audit-dir\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.565787 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55b6dd11-c219-4b24-90eb-dbc096a67835-config\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.566062 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.566365 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-serving-cert\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.566564 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.567709 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/48a0fa25-6b2d-4668-b8e5-824912077f19-encryption-config\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.567908 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/48a0fa25-6b2d-4668-b8e5-824912077f19-image-import-ca\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.567994 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.568019 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-etcd-client\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.568315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c2f00a8-c3ce-4957-9093-d8c2cce49992-serving-cert\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.568616 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.568888 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.569365 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-serving-cert\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.569489 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/7e94e6e0-b16d-462f-b791-ba20acdcb809-encryption-config\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.569688 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.570066 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.570116 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.570208 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rx7tc"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.570888 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.571084 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.571422 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555646-wqpkb"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.571768 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.572468 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.572500 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.572515 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4230f869-9456-44a1-87b3-342fc8c18ed7-serving-cert\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.572714 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.573269 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qs58t"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.573305 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.573367 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.573519 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.574138 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-service-ca-bundle\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.574486 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bfn44"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.574508 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.574520 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.574531 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-z95pp"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.574544 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.574991 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-serving-cert\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.575138 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.575190 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.575234 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-rxml7"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.575268 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.575426 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.576330 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-c9xl5"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.576626 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.585480 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.585519 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bsvxw"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.585531 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.589750 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.589867 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c30a7ad-f92b-445a-9201-fe55f247cf41-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.590304 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4c2f00a8-c3ce-4957-9093-d8c2cce49992-etcd-client\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.590397 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/55b6dd11-c219-4b24-90eb-dbc096a67835-machine-approver-tls\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.590792 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.592286 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.593515 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.595333 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9zg78"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.599371 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2wsbc"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.599984 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-serving-cert\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.600609 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.604478 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5db62dca-ba86-4ca4-861e-003d09e5ac0f-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-hwwdz\" (UID: \"5db62dca-ba86-4ca4-861e-003d09e5ac0f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.606320 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.610169 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.611392 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-st67x"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.620279 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.621632 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q49l5"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.622372 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee39302-0316-4481-871e-538ffd31a507-config\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623071 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-metrics-certs\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623198 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfrhq\" (UniqueName: \"kubernetes.io/projected/1dc82997-2782-4e9e-a293-956fcb96acde-kube-api-access-xfrhq\") pod \"dns-operator-744455d44c-9zg78\" (UID: \"1dc82997-2782-4e9e-a293-956fcb96acde\") " pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623242 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ee39302-0316-4481-871e-538ffd31a507-trusted-ca\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623266 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1dc82997-2782-4e9e-a293-956fcb96acde-metrics-tls\") pod \"dns-operator-744455d44c-9zg78\" (UID: \"1dc82997-2782-4e9e-a293-956fcb96acde\") " pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623297 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623344 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct49b\" (UniqueName: \"kubernetes.io/projected/e379fe1d-7780-4f17-8df8-f74f3dddbc23-kube-api-access-ct49b\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623374 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-config\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623395 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-stats-auth\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623835 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ee39302-0316-4481-871e-538ffd31a507-serving-cert\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623849 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ee39302-0316-4481-871e-538ffd31a507-config\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623946 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjftj\" (UniqueName: \"kubernetes.io/projected/7ee39302-0316-4481-871e-538ffd31a507-kube-api-access-tjftj\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.623975 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-default-certificate\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.624035 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e379fe1d-7780-4f17-8df8-f74f3dddbc23-service-ca-bundle\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.626505 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7ee39302-0316-4481-871e-538ffd31a507-trusted-ca\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.626628 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.628245 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1dc82997-2782-4e9e-a293-956fcb96acde-metrics-tls\") pod \"dns-operator-744455d44c-9zg78\" (UID: \"1dc82997-2782-4e9e-a293-956fcb96acde\") " pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.628811 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.633203 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s9tvm"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.633863 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ee39302-0316-4481-871e-538ffd31a507-serving-cert\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.635175 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7rmsn"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.638873 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vb9qx"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.640935 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.641707 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c68kr"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.644147 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rx7tc"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.645713 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.648130 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.650103 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h95sl"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.651168 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.652457 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.653938 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.654837 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.656390 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rxml7"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.657421 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.659595 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g6fzt"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.660664 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.660636 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.661692 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.662696 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-drmq2"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.663917 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.664078 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-pd4lm"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.665003 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.665017 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555646-wqpkb"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.666065 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pd4lm"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.667416 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-c9xl5"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.668798 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f"] Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.687117 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.701235 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.720702 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.741808 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.769817 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.780904 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.801509 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.821308 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.841009 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.860358 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.880741 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.900580 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.921049 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.940547 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.961217 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 12 18:06:22 crc kubenswrapper[4926]: I0312 18:06:22.982091 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.001643 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.021187 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.041657 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.061158 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.081088 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.101292 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.120989 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.141374 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.148345 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.161706 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.167158 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-config\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.201912 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.208208 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-metrics-certs\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.221240 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.230158 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-default-certificate\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.241758 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.250732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e379fe1d-7780-4f17-8df8-f74f3dddbc23-stats-auth\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.261040 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.281628 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.301214 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.307106 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e379fe1d-7780-4f17-8df8-f74f3dddbc23-service-ca-bundle\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.322584 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.341233 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.361183 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.381785 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.454930 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.457614 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.458568 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.460775 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.481152 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.499859 4926 request.go:700] Waited for 1.011481135s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serving-cert&limit=500&resourceVersion=0 Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.501810 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.521119 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.549375 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.561254 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.580683 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.600141 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.620814 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.641384 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.660879 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.682005 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.701755 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.721590 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.742125 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.761953 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.801333 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.846777 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bh26\" (UniqueName: \"kubernetes.io/projected/49e5e304-df7c-434b-8b17-f520e9bb7d52-kube-api-access-6bh26\") pod \"machine-api-operator-5694c8668f-bfn44\" (UID: \"49e5e304-df7c-434b-8b17-f520e9bb7d52\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.861667 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t46vq\" (UniqueName: \"kubernetes.io/projected/270031fa-3d83-4edf-bb5d-19ce9e1a693d-kube-api-access-t46vq\") pod \"console-f9d7485db-vb9qx\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.895586 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88hg4\" (UniqueName: \"kubernetes.io/projected/eedd886d-5443-47e1-afbf-5aff90067f3b-kube-api-access-88hg4\") pod \"oauth-openshift-558db77b4-7rmsn\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.908024 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62f9n\" (UniqueName: \"kubernetes.io/projected/5db62dca-ba86-4ca4-861e-003d09e5ac0f-kube-api-access-62f9n\") pod \"cluster-samples-operator-665b6dd947-hwwdz\" (UID: \"5db62dca-ba86-4ca4-861e-003d09e5ac0f\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.919050 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqbzk\" (UniqueName: \"kubernetes.io/projected/64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a-kube-api-access-hqbzk\") pod \"authentication-operator-69f744f599-qs58t\" (UID: \"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.936334 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbgdh\" (UniqueName: \"kubernetes.io/projected/55b6dd11-c219-4b24-90eb-dbc096a67835-kube-api-access-jbgdh\") pod \"machine-approver-56656f9798-qbxxz\" (UID: \"55b6dd11-c219-4b24-90eb-dbc096a67835\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.955004 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w622m\" (UniqueName: \"kubernetes.io/projected/4c2f00a8-c3ce-4957-9093-d8c2cce49992-kube-api-access-w622m\") pod \"etcd-operator-b45778765-2wsbc\" (UID: \"4c2f00a8-c3ce-4957-9093-d8c2cce49992\") " pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.976542 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5ctp\" (UniqueName: \"kubernetes.io/projected/a5e9834d-5aeb-4154-ae70-c2b6b07c9eca-kube-api-access-n5ctp\") pod \"openshift-apiserver-operator-796bbdcf4f-q2smg\" (UID: \"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.994633 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7dxn\" (UniqueName: \"kubernetes.io/projected/8c30a7ad-f92b-445a-9201-fe55f247cf41-kube-api-access-c7dxn\") pod \"openshift-controller-manager-operator-756b6f6bc6-x7fm2\" (UID: \"8c30a7ad-f92b-445a-9201-fe55f247cf41\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:23 crc kubenswrapper[4926]: I0312 18:06:23.998048 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.018528 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.019125 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj7xf\" (UniqueName: \"kubernetes.io/projected/7e94e6e0-b16d-462f-b791-ba20acdcb809-kube-api-access-fj7xf\") pod \"apiserver-7bbb656c7d-9bzkt\" (UID: \"7e94e6e0-b16d-462f-b791-ba20acdcb809\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.049223 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snkcw\" (UniqueName: \"kubernetes.io/projected/48a0fa25-6b2d-4668-b8e5-824912077f19-kube-api-access-snkcw\") pod \"apiserver-76f77b778f-z95pp\" (UID: \"48a0fa25-6b2d-4668-b8e5-824912077f19\") " pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.062550 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.063885 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28x4l\" (UniqueName: \"kubernetes.io/projected/4230f869-9456-44a1-87b3-342fc8c18ed7-kube-api-access-28x4l\") pod \"route-controller-manager-6576b87f9c-svrln\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.081990 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.101414 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.118684 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.121503 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.124686 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.127375 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.141767 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.141832 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.147802 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.155605 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.162619 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.163296 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.171583 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.182004 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.201038 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.213823 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.222412 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.233553 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.241685 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.265523 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.280992 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.302205 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.302497 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-bfn44"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.330172 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.344396 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.361035 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.381105 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.413975 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.435916 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.457359 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.461186 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vb9qx"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.474084 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.480769 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.494756 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vb9qx" event={"ID":"270031fa-3d83-4edf-bb5d-19ce9e1a693d","Type":"ContainerStarted","Data":"8b8722347bac291c780b1fdd439b9d8743876830528929d7bd120c0953d14111"} Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.494814 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" event={"ID":"55b6dd11-c219-4b24-90eb-dbc096a67835","Type":"ContainerStarted","Data":"75436a3df44b0e6d0f107c0ad8e4ab6be3973a98a11e68d34310dc5eea08705c"} Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.494828 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" event={"ID":"49e5e304-df7c-434b-8b17-f520e9bb7d52","Type":"ContainerStarted","Data":"76a0ae2a34847f15c5c8c0585f4a7daf639c1bba56be8e5de82c8ced6f6edcac"} Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.500337 4926 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.520530 4926 request.go:700] Waited for 1.929203658s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/hostpath-provisioner/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.523345 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.559161 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfrhq\" (UniqueName: \"kubernetes.io/projected/1dc82997-2782-4e9e-a293-956fcb96acde-kube-api-access-xfrhq\") pod \"dns-operator-744455d44c-9zg78\" (UID: \"1dc82997-2782-4e9e-a293-956fcb96acde\") " pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.596292 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjftj\" (UniqueName: \"kubernetes.io/projected/7ee39302-0316-4481-871e-538ffd31a507-kube-api-access-tjftj\") pod \"console-operator-58897d9998-s9tvm\" (UID: \"7ee39302-0316-4481-871e-538ffd31a507\") " pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.604268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct49b\" (UniqueName: \"kubernetes.io/projected/e379fe1d-7780-4f17-8df8-f74f3dddbc23-kube-api-access-ct49b\") pod \"router-default-5444994796-wndvq\" (UID: \"e379fe1d-7780-4f17-8df8-f74f3dddbc23\") " pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.615299 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5cae3622-73dc-43ee-9e5c-eb6c67e37c1e-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6hxxs\" (UID: \"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.620467 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.626720 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2"] Mar 12 18:06:24 crc kubenswrapper[4926]: W0312 18:06:24.637021 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c30a7ad_f92b_445a_9201_fe55f247cf41.slice/crio-b55657cb73acd5feec151622d460258d2ed79fb0c98212b3ade5d94e9bbe6b62 WatchSource:0}: Error finding container b55657cb73acd5feec151622d460258d2ed79fb0c98212b3ade5d94e9bbe6b62: Status 404 returned error can't find the container with id b55657cb73acd5feec151622d460258d2ed79fb0c98212b3ade5d94e9bbe6b62 Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.640959 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.660235 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.682877 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.698861 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.699586 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.701595 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.706824 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-z95pp"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.707868 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qs58t"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.713240 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-2wsbc"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.716200 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7rmsn"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.720598 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.731917 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.740568 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.776318 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.828579 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.856515 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.856743 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.858862 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.882926 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-bound-sa-token\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.882955 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wq55\" (UniqueName: \"kubernetes.io/projected/f6ad9439-d12c-4987-840a-002975ba1498-kube-api-access-4wq55\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.882972 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883007 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-serving-cert\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883022 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f68a2b5d-036a-43ee-a9f2-dd94d9f51d51-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bsvxw\" (UID: \"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883046 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-srv-cert\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883061 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4867106-501e-4408-af0b-7790cfc45a24-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883098 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb73641e-b156-470d-be1a-52c11f2efdf6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883119 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4wwg\" (UniqueName: \"kubernetes.io/projected/fb73641e-b156-470d-be1a-52c11f2efdf6-kube-api-access-d4wwg\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883135 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5212edbb-b23e-44b8-aab1-77dece6d1415-serving-cert\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883170 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5212edbb-b23e-44b8-aab1-77dece6d1415-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883186 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlg4x\" (UniqueName: \"kubernetes.io/projected/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-kube-api-access-nlg4x\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883209 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883223 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-config\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883248 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zvfj\" (UniqueName: \"kubernetes.io/projected/f68a2b5d-036a-43ee-a9f2-dd94d9f51d51-kube-api-access-8zvfj\") pod \"multus-admission-controller-857f4d67dd-bsvxw\" (UID: \"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883271 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36a0d781-e388-41ae-878e-05a69d81c83e-config\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883292 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883317 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97b0faa2-bcb2-417e-9065-3156860a8644-installation-pull-secrets\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883333 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883367 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6ad9439-d12c-4987-840a-002975ba1498-profile-collector-cert\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883388 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883403 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb73641e-b156-470d-be1a-52c11f2efdf6-trusted-ca\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883418 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931c621-236d-4a2f-9c96-2c29483f19db-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883458 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-trusted-ca\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883474 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-registry-tls\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883488 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97b0faa2-bcb2-417e-9065-3156860a8644-ca-trust-extracted\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883537 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a931c621-236d-4a2f-9c96-2c29483f19db-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883554 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2x5w\" (UniqueName: \"kubernetes.io/projected/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-kube-api-access-m2x5w\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883592 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdq7x\" (UniqueName: \"kubernetes.io/projected/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-kube-api-access-vdq7x\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883606 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a0d781-e388-41ae-878e-05a69d81c83e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883638 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4867106-501e-4408-af0b-7790cfc45a24-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883653 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36a0d781-e388-41ae-878e-05a69d81c83e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883671 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a931c621-236d-4a2f-9c96-2c29483f19db-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883688 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4gdc\" (UniqueName: \"kubernetes.io/projected/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-kube-api-access-x4gdc\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d073c88-4608-4594-9feb-f1093455368d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sgl9z\" (UID: \"5d073c88-4608-4594-9feb-f1093455368d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883721 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6ad9439-d12c-4987-840a-002975ba1498-srv-cert\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883735 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htq94\" (UniqueName: \"kubernetes.io/projected/f4867106-501e-4408-af0b-7790cfc45a24-kube-api-access-htq94\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883750 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4867106-501e-4408-af0b-7790cfc45a24-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883785 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpjs6\" (UniqueName: \"kubernetes.io/projected/439a0925-eff6-4da1-b65c-de0623b9daaa-kube-api-access-wpjs6\") pod \"migrator-59844c95c7-sbn5h\" (UID: \"439a0925-eff6-4da1-b65c-de0623b9daaa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883801 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrh7w\" (UniqueName: \"kubernetes.io/projected/5d073c88-4608-4594-9feb-f1093455368d-kube-api-access-nrh7w\") pod \"control-plane-machine-set-operator-78cbb6b69f-sgl9z\" (UID: \"5d073c88-4608-4594-9feb-f1093455368d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883814 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-profile-collector-cert\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883828 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltlt7\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-kube-api-access-ltlt7\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883861 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883876 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883892 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bqbj\" (UniqueName: \"kubernetes.io/projected/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-kube-api-access-5bqbj\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883909 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-client-ca\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883931 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-registry-certificates\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883973 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-proxy-tls\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.883989 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blwnd\" (UniqueName: \"kubernetes.io/projected/06394469-defd-4710-90a2-b6c395c00d4f-kube-api-access-blwnd\") pod \"downloads-7954f5f757-st67x\" (UID: \"06394469-defd-4710-90a2-b6c395c00d4f\") " pod="openshift-console/downloads-7954f5f757-st67x" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.884004 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-images\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.884018 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fb73641e-b156-470d-be1a-52c11f2efdf6-metrics-tls\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.884031 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92cqr\" (UniqueName: \"kubernetes.io/projected/5212edbb-b23e-44b8-aab1-77dece6d1415-kube-api-access-92cqr\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:24 crc kubenswrapper[4926]: E0312 18:06:24.886785 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:25.386769638 +0000 UTC m=+225.755395971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:24 crc kubenswrapper[4926]: W0312 18:06:24.915842 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4230f869_9456_44a1_87b3_342fc8c18ed7.slice/crio-01692fb622a525e9c42adef3bc2dd9239f99fa47f15b51038978d9cd4978f5eb WatchSource:0}: Error finding container 01692fb622a525e9c42adef3bc2dd9239f99fa47f15b51038978d9cd4978f5eb: Status 404 returned error can't find the container with id 01692fb622a525e9c42adef3bc2dd9239f99fa47f15b51038978d9cd4978f5eb Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.944204 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-s9tvm"] Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.984526 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-srv-cert\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985129 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4867106-501e-4408-af0b-7790cfc45a24-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985153 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-webhook-cert\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxlnr\" (UniqueName: \"kubernetes.io/projected/416787ca-b62c-4fd5-84c5-57be59317faa-kube-api-access-rxlnr\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985214 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb73641e-b156-470d-be1a-52c11f2efdf6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985233 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-config-volume\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985250 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4wwg\" (UniqueName: \"kubernetes.io/projected/fb73641e-b156-470d-be1a-52c11f2efdf6-kube-api-access-d4wwg\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985268 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5212edbb-b23e-44b8-aab1-77dece6d1415-serving-cert\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985285 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nzfp\" (UniqueName: \"kubernetes.io/projected/721fd54d-55f5-4477-a7b8-283a700c8c47-kube-api-access-8nzfp\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4edd042a-f910-49c2-9220-56a6e79b04dc-config-volume\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985319 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlg4x\" (UniqueName: \"kubernetes.io/projected/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-kube-api-access-nlg4x\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985335 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-plugins-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985358 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5212edbb-b23e-44b8-aab1-77dece6d1415-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985374 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/721fd54d-55f5-4477-a7b8-283a700c8c47-signing-cabundle\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985391 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-config\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985430 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zvfj\" (UniqueName: \"kubernetes.io/projected/f68a2b5d-036a-43ee-a9f2-dd94d9f51d51-kube-api-access-8zvfj\") pod \"multus-admission-controller-857f4d67dd-bsvxw\" (UID: \"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985462 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36a0d781-e388-41ae-878e-05a69d81c83e-config\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985491 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq9lz\" (UniqueName: \"kubernetes.io/projected/4edd042a-f910-49c2-9220-56a6e79b04dc-kube-api-access-hq9lz\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985512 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f14f5e4-148e-4282-9243-89c96906048a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jfb58\" (UID: \"0f14f5e4-148e-4282-9243-89c96906048a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6gw4\" (UniqueName: \"kubernetes.io/projected/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-kube-api-access-b6gw4\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985559 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtltr\" (UniqueName: \"kubernetes.io/projected/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-kube-api-access-rtltr\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985583 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-socket-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985597 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c574438a-86a5-4a9f-aff3-47bf920cbabd-cert\") pod \"ingress-canary-pd4lm\" (UID: \"c574438a-86a5-4a9f-aff3-47bf920cbabd\") " pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985615 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97b0faa2-bcb2-417e-9065-3156860a8644-installation-pull-secrets\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985637 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d8f4\" (UniqueName: \"kubernetes.io/projected/c574438a-86a5-4a9f-aff3-47bf920cbabd-kube-api-access-2d8f4\") pod \"ingress-canary-pd4lm\" (UID: \"c574438a-86a5-4a9f-aff3-47bf920cbabd\") " pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985678 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6ad9439-d12c-4987-840a-002975ba1498-profile-collector-cert\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985705 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/721fd54d-55f5-4477-a7b8-283a700c8c47-signing-key\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985720 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb73641e-b156-470d-be1a-52c11f2efdf6-trusted-ca\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985758 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931c621-236d-4a2f-9c96-2c29483f19db-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985784 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-trusted-ca\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985799 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-registration-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985814 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4edd042a-f910-49c2-9220-56a6e79b04dc-secret-volume\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985829 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-registry-tls\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985843 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97b0faa2-bcb2-417e-9065-3156860a8644-ca-trust-extracted\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a931c621-236d-4a2f-9c96-2c29483f19db-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985908 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-apiservice-cert\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985941 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2x5w\" (UniqueName: \"kubernetes.io/projected/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-kube-api-access-m2x5w\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985957 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xbdk\" (UniqueName: \"kubernetes.io/projected/8acca805-2bf8-4dcb-a036-c4732084210e-kube-api-access-4xbdk\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.985989 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-mountpoint-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986003 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00918e77-d6a1-4d1b-9986-37f3e648b322-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986020 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdq7x\" (UniqueName: \"kubernetes.io/projected/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-kube-api-access-vdq7x\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986035 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a0d781-e388-41ae-878e-05a69d81c83e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986053 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4867106-501e-4408-af0b-7790cfc45a24-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986068 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wx8l\" (UniqueName: \"kubernetes.io/projected/00918e77-d6a1-4d1b-9986-37f3e648b322-kube-api-access-2wx8l\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986083 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36a0d781-e388-41ae-878e-05a69d81c83e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986098 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-metrics-tls\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986114 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a931c621-236d-4a2f-9c96-2c29483f19db-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986146 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4gdc\" (UniqueName: \"kubernetes.io/projected/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-kube-api-access-x4gdc\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986172 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d073c88-4608-4594-9feb-f1093455368d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sgl9z\" (UID: \"5d073c88-4608-4594-9feb-f1093455368d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986189 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrq7x\" (UniqueName: \"kubernetes.io/projected/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-kube-api-access-nrq7x\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986207 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6ad9439-d12c-4987-840a-002975ba1498-srv-cert\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986222 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htq94\" (UniqueName: \"kubernetes.io/projected/f4867106-501e-4408-af0b-7790cfc45a24-kube-api-access-htq94\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986245 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4867106-501e-4408-af0b-7790cfc45a24-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986261 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00918e77-d6a1-4d1b-9986-37f3e648b322-proxy-tls\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986275 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pshhr\" (UniqueName: \"kubernetes.io/projected/0f14f5e4-148e-4282-9243-89c96906048a-kube-api-access-pshhr\") pod \"package-server-manager-789f6589d5-jfb58\" (UID: \"0f14f5e4-148e-4282-9243-89c96906048a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986293 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpjs6\" (UniqueName: \"kubernetes.io/projected/439a0925-eff6-4da1-b65c-de0623b9daaa-kube-api-access-wpjs6\") pod \"migrator-59844c95c7-sbn5h\" (UID: \"439a0925-eff6-4da1-b65c-de0623b9daaa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986310 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/416787ca-b62c-4fd5-84c5-57be59317faa-serving-cert\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986326 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416787ca-b62c-4fd5-84c5-57be59317faa-config\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986352 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrh7w\" (UniqueName: \"kubernetes.io/projected/5d073c88-4608-4594-9feb-f1093455368d-kube-api-access-nrh7w\") pod \"control-plane-machine-set-operator-78cbb6b69f-sgl9z\" (UID: \"5d073c88-4608-4594-9feb-f1093455368d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986368 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-profile-collector-cert\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986401 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltlt7\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-kube-api-access-ltlt7\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986419 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-node-bootstrap-token\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986830 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986851 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-client-ca\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986883 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bqbj\" (UniqueName: \"kubernetes.io/projected/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-kube-api-access-5bqbj\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986899 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-tmpfs\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986915 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-registry-certificates\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986930 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl7pp\" (UniqueName: \"kubernetes.io/projected/d68160cf-4e6c-4294-bfdc-4acb74637ecb-kube-api-access-tl7pp\") pod \"auto-csr-approver-29555646-wqpkb\" (UID: \"d68160cf-4e6c-4294-bfdc-4acb74637ecb\") " pod="openshift-infra/auto-csr-approver-29555646-wqpkb" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986971 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-proxy-tls\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.986987 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blwnd\" (UniqueName: \"kubernetes.io/projected/06394469-defd-4710-90a2-b6c395c00d4f-kube-api-access-blwnd\") pod \"downloads-7954f5f757-st67x\" (UID: \"06394469-defd-4710-90a2-b6c395c00d4f\") " pod="openshift-console/downloads-7954f5f757-st67x" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987004 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-images\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987039 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fb73641e-b156-470d-be1a-52c11f2efdf6-metrics-tls\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987055 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92cqr\" (UniqueName: \"kubernetes.io/projected/5212edbb-b23e-44b8-aab1-77dece6d1415-kube-api-access-92cqr\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987083 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-bound-sa-token\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987097 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-certs\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987141 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wq55\" (UniqueName: \"kubernetes.io/projected/f6ad9439-d12c-4987-840a-002975ba1498-kube-api-access-4wq55\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987157 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-csi-data-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-serving-cert\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:24 crc kubenswrapper[4926]: I0312 18:06:24.987263 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f68a2b5d-036a-43ee-a9f2-dd94d9f51d51-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bsvxw\" (UID: \"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:24.996497 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:25.496463123 +0000 UTC m=+225.865089466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:24.998184 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f68a2b5d-036a-43ee-a9f2-dd94d9f51d51-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-bsvxw\" (UID: \"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.000147 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-srv-cert\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.003690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.004651 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.009028 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97b0faa2-bcb2-417e-9065-3156860a8644-ca-trust-extracted\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.011492 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97b0faa2-bcb2-417e-9065-3156860a8644-installation-pull-secrets\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.011803 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4867106-501e-4408-af0b-7790cfc45a24-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.012619 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a931c621-236d-4a2f-9c96-2c29483f19db-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.013495 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.013525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36a0d781-e388-41ae-878e-05a69d81c83e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.013891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.014166 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-trusted-ca\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.014424 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36a0d781-e388-41ae-878e-05a69d81c83e-config\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.014571 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931c621-236d-4a2f-9c96-2c29483f19db-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.015287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-client-ca\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.015419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5212edbb-b23e-44b8-aab1-77dece6d1415-available-featuregates\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.015592 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5212edbb-b23e-44b8-aab1-77dece6d1415-serving-cert\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.015749 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb73641e-b156-470d-be1a-52c11f2efdf6-trusted-ca\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.016181 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-registry-certificates\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.016717 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-auth-proxy-config\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.019209 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-config\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.019418 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-images\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.021533 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5d073c88-4608-4594-9feb-f1093455368d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-sgl9z\" (UID: \"5d073c88-4608-4594-9feb-f1093455368d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.021996 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-profile-collector-cert\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.024385 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9zg78"] Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.029677 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4867106-501e-4408-af0b-7790cfc45a24-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.029710 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f6ad9439-d12c-4987-840a-002975ba1498-srv-cert\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.029849 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-serving-cert\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.030156 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-proxy-tls\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.031054 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f6ad9439-d12c-4987-840a-002975ba1498-profile-collector-cert\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.032828 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.038191 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-registry-tls\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.038645 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fb73641e-b156-470d-be1a-52c11f2efdf6-metrics-tls\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.039381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92cqr\" (UniqueName: \"kubernetes.io/projected/5212edbb-b23e-44b8-aab1-77dece6d1415-kube-api-access-92cqr\") pod \"openshift-config-operator-7777fb866f-h95sl\" (UID: \"5212edbb-b23e-44b8-aab1-77dece6d1415\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.058310 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdq7x\" (UniqueName: \"kubernetes.io/projected/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-kube-api-access-vdq7x\") pod \"marketplace-operator-79b997595-c68kr\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.079533 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36a0d781-e388-41ae-878e-05a69d81c83e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sj8r7\" (UID: \"36a0d781-e388-41ae-878e-05a69d81c83e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.082987 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089585 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-webhook-cert\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxlnr\" (UniqueName: \"kubernetes.io/projected/416787ca-b62c-4fd5-84c5-57be59317faa-kube-api-access-rxlnr\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089646 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-config-volume\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089670 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nzfp\" (UniqueName: \"kubernetes.io/projected/721fd54d-55f5-4477-a7b8-283a700c8c47-kube-api-access-8nzfp\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4edd042a-f910-49c2-9220-56a6e79b04dc-config-volume\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089708 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-plugins-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089726 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/721fd54d-55f5-4477-a7b8-283a700c8c47-signing-cabundle\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089755 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq9lz\" (UniqueName: \"kubernetes.io/projected/4edd042a-f910-49c2-9220-56a6e79b04dc-kube-api-access-hq9lz\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089774 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089789 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f14f5e4-148e-4282-9243-89c96906048a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jfb58\" (UID: \"0f14f5e4-148e-4282-9243-89c96906048a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089806 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6gw4\" (UniqueName: \"kubernetes.io/projected/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-kube-api-access-b6gw4\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtltr\" (UniqueName: \"kubernetes.io/projected/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-kube-api-access-rtltr\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089836 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-socket-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089854 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c574438a-86a5-4a9f-aff3-47bf920cbabd-cert\") pod \"ingress-canary-pd4lm\" (UID: \"c574438a-86a5-4a9f-aff3-47bf920cbabd\") " pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089871 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d8f4\" (UniqueName: \"kubernetes.io/projected/c574438a-86a5-4a9f-aff3-47bf920cbabd-kube-api-access-2d8f4\") pod \"ingress-canary-pd4lm\" (UID: \"c574438a-86a5-4a9f-aff3-47bf920cbabd\") " pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089887 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/721fd54d-55f5-4477-a7b8-283a700c8c47-signing-key\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-registration-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089917 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4edd042a-f910-49c2-9220-56a6e79b04dc-secret-volume\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089951 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-apiservice-cert\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089974 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xbdk\" (UniqueName: \"kubernetes.io/projected/8acca805-2bf8-4dcb-a036-c4732084210e-kube-api-access-4xbdk\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.089994 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-mountpoint-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090014 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00918e77-d6a1-4d1b-9986-37f3e648b322-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090047 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wx8l\" (UniqueName: \"kubernetes.io/projected/00918e77-d6a1-4d1b-9986-37f3e648b322-kube-api-access-2wx8l\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-metrics-tls\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090100 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrq7x\" (UniqueName: \"kubernetes.io/projected/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-kube-api-access-nrq7x\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090120 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00918e77-d6a1-4d1b-9986-37f3e648b322-proxy-tls\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090134 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pshhr\" (UniqueName: \"kubernetes.io/projected/0f14f5e4-148e-4282-9243-89c96906048a-kube-api-access-pshhr\") pod \"package-server-manager-789f6589d5-jfb58\" (UID: \"0f14f5e4-148e-4282-9243-89c96906048a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090160 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/416787ca-b62c-4fd5-84c5-57be59317faa-serving-cert\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090179 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416787ca-b62c-4fd5-84c5-57be59317faa-config\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090202 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-node-bootstrap-token\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090233 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-tmpfs\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090248 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl7pp\" (UniqueName: \"kubernetes.io/projected/d68160cf-4e6c-4294-bfdc-4acb74637ecb-kube-api-access-tl7pp\") pod \"auto-csr-approver-29555646-wqpkb\" (UID: \"d68160cf-4e6c-4294-bfdc-4acb74637ecb\") " pod="openshift-infra/auto-csr-approver-29555646-wqpkb" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090281 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-certs\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090298 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-csi-data-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.090409 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-csi-data-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.091135 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-config-volume\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.091185 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4edd042a-f910-49c2-9220-56a6e79b04dc-config-volume\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.093969 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:25.593957707 +0000 UTC m=+225.962584040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.094392 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-plugins-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.094910 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-mountpoint-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.095191 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/721fd54d-55f5-4477-a7b8-283a700c8c47-signing-cabundle\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.095284 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-registration-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.095320 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8acca805-2bf8-4dcb-a036-c4732084210e-socket-dir\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.096009 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/00918e77-d6a1-4d1b-9986-37f3e648b322-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.097075 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs"] Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.097085 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416787ca-b62c-4fd5-84c5-57be59317faa-config\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.097353 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-tmpfs\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.097388 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c574438a-86a5-4a9f-aff3-47bf920cbabd-cert\") pod \"ingress-canary-pd4lm\" (UID: \"c574438a-86a5-4a9f-aff3-47bf920cbabd\") " pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.099302 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/721fd54d-55f5-4477-a7b8-283a700c8c47-signing-key\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.099966 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-webhook-cert\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.100690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/00918e77-d6a1-4d1b-9986-37f3e648b322-proxy-tls\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.101321 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4edd042a-f910-49c2-9220-56a6e79b04dc-secret-volume\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.104239 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-apiservice-cert\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.104424 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f14f5e4-148e-4282-9243-89c96906048a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jfb58\" (UID: \"0f14f5e4-148e-4282-9243-89c96906048a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.105997 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4867106-501e-4408-af0b-7790cfc45a24-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.106268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-node-bootstrap-token\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.110904 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/416787ca-b62c-4fd5-84c5-57be59317faa-serving-cert\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.111013 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-certs\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.114513 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-metrics-tls\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.124035 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-bound-sa-token\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.143217 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a931c621-236d-4a2f-9c96-2c29483f19db-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-b9s9h\" (UID: \"a931c621-236d-4a2f-9c96-2c29483f19db\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.160326 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4gdc\" (UniqueName: \"kubernetes.io/projected/f54babbf-5b93-4ef7-99d7-0cfe05e921c5-kube-api-access-x4gdc\") pod \"kube-storage-version-migrator-operator-b67b599dd-snpzd\" (UID: \"f54babbf-5b93-4ef7-99d7-0cfe05e921c5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.186758 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.189958 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.200682 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.201456 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:25.70142186 +0000 UTC m=+226.070048193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.206330 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.217348 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wq55\" (UniqueName: \"kubernetes.io/projected/f6ad9439-d12c-4987-840a-002975ba1498-kube-api-access-4wq55\") pod \"olm-operator-6b444d44fb-x4kkc\" (UID: \"f6ad9439-d12c-4987-840a-002975ba1498\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.230780 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htq94\" (UniqueName: \"kubernetes.io/projected/f4867106-501e-4408-af0b-7790cfc45a24-kube-api-access-htq94\") pod \"cluster-image-registry-operator-dc59b4c8b-xxzt9\" (UID: \"f4867106-501e-4408-af0b-7790cfc45a24\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.236746 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2x5w\" (UniqueName: \"kubernetes.io/projected/b99fa2a7-b47e-433c-b6d5-7c1306a79cde-kube-api-access-m2x5w\") pod \"catalog-operator-68c6474976-rdjk9\" (UID: \"b99fa2a7-b47e-433c-b6d5-7c1306a79cde\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.257096 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb73641e-b156-470d-be1a-52c11f2efdf6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.275814 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4wwg\" (UniqueName: \"kubernetes.io/projected/fb73641e-b156-470d-be1a-52c11f2efdf6-kube-api-access-d4wwg\") pod \"ingress-operator-5b745b69d9-h5mf8\" (UID: \"fb73641e-b156-470d-be1a-52c11f2efdf6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.301021 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zvfj\" (UniqueName: \"kubernetes.io/projected/f68a2b5d-036a-43ee-a9f2-dd94d9f51d51-kube-api-access-8zvfj\") pod \"multus-admission-controller-857f4d67dd-bsvxw\" (UID: \"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.302028 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-h95sl"] Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.302970 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.303331 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:25.803319536 +0000 UTC m=+226.171945869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.318933 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpjs6\" (UniqueName: \"kubernetes.io/projected/439a0925-eff6-4da1-b65c-de0623b9daaa-kube-api-access-wpjs6\") pod \"migrator-59844c95c7-sbn5h\" (UID: \"439a0925-eff6-4da1-b65c-de0623b9daaa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.348877 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlg4x\" (UniqueName: \"kubernetes.io/projected/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-kube-api-access-nlg4x\") pod \"controller-manager-879f6c89f-q49l5\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.366247 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bqbj\" (UniqueName: \"kubernetes.io/projected/2e2b819b-3df8-4d51-b6d0-b30e33f1ceba-kube-api-access-5bqbj\") pod \"machine-config-operator-74547568cd-bmp6d\" (UID: \"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.388822 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrh7w\" (UniqueName: \"kubernetes.io/projected/5d073c88-4608-4594-9feb-f1093455368d-kube-api-access-nrh7w\") pod \"control-plane-machine-set-operator-78cbb6b69f-sgl9z\" (UID: \"5d073c88-4608-4594-9feb-f1093455368d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.399066 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.402291 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltlt7\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-kube-api-access-ltlt7\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.406325 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.406792 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:25.906774808 +0000 UTC m=+226.275401141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.407298 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.412520 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.422000 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.423895 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blwnd\" (UniqueName: \"kubernetes.io/projected/06394469-defd-4710-90a2-b6c395c00d4f-kube-api-access-blwnd\") pod \"downloads-7954f5f757-st67x\" (UID: \"06394469-defd-4710-90a2-b6c395c00d4f\") " pod="openshift-console/downloads-7954f5f757-st67x" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.428517 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.436202 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.436841 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nzfp\" (UniqueName: \"kubernetes.io/projected/721fd54d-55f5-4477-a7b8-283a700c8c47-kube-api-access-8nzfp\") pod \"service-ca-9c57cc56f-rx7tc\" (UID: \"721fd54d-55f5-4477-a7b8-283a700c8c47\") " pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.444741 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.455748 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wx8l\" (UniqueName: \"kubernetes.io/projected/00918e77-d6a1-4d1b-9986-37f3e648b322-kube-api-access-2wx8l\") pod \"machine-config-controller-84d6567774-dnk9m\" (UID: \"00918e77-d6a1-4d1b-9986-37f3e648b322\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.465889 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.478747 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxlnr\" (UniqueName: \"kubernetes.io/projected/416787ca-b62c-4fd5-84c5-57be59317faa-kube-api-access-rxlnr\") pod \"service-ca-operator-777779d784-z4g7g\" (UID: \"416787ca-b62c-4fd5-84c5-57be59317faa\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.487892 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.499414 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.507390 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d8f4\" (UniqueName: \"kubernetes.io/projected/c574438a-86a5-4a9f-aff3-47bf920cbabd-kube-api-access-2d8f4\") pod \"ingress-canary-pd4lm\" (UID: \"c574438a-86a5-4a9f-aff3-47bf920cbabd\") " pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.516985 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xbdk\" (UniqueName: \"kubernetes.io/projected/8acca805-2bf8-4dcb-a036-c4732084210e-kube-api-access-4xbdk\") pod \"csi-hostpathplugin-c9xl5\" (UID: \"8acca805-2bf8-4dcb-a036-c4732084210e\") " pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.523868 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.525556 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.525886 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.025870754 +0000 UTC m=+226.394497087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.533689 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.544059 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" event={"ID":"1dc82997-2782-4e9e-a293-956fcb96acde","Type":"ContainerStarted","Data":"45644e365c4c4c529f9a87c6adeb38c71d01569ba997e441f62dc7542cad62f8"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.544105 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" event={"ID":"1dc82997-2782-4e9e-a293-956fcb96acde","Type":"ContainerStarted","Data":"a481d30dbd54dc572c52e184a6bde6b513a6eb7b1a50a62d5566922da33a686d"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.546786 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtltr\" (UniqueName: \"kubernetes.io/projected/4c9eaabb-c0c2-4f71-82eb-7494d0a37075-kube-api-access-rtltr\") pod \"dns-default-rxml7\" (UID: \"4c9eaabb-c0c2-4f71-82eb-7494d0a37075\") " pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.561571 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" event={"ID":"5db62dca-ba86-4ca4-861e-003d09e5ac0f","Type":"ContainerStarted","Data":"d6d9c79e93d5134e8d839538f09145ae2bd4d77bbe3f38c3f1f15cf37a650938"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.561616 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" event={"ID":"5db62dca-ba86-4ca4-861e-003d09e5ac0f","Type":"ContainerStarted","Data":"0182f53fcf7be43817f7e2563d3881425c615ed6aa5302372ea5d8c19324879d"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.561629 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" event={"ID":"5db62dca-ba86-4ca4-861e-003d09e5ac0f","Type":"ContainerStarted","Data":"ad09cbb8919dc168feec360f061267f9c6dab1c238365c26b99a91802e4a61ea"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.578651 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6gw4\" (UniqueName: \"kubernetes.io/projected/6aa7a74a-b3d4-47df-a1b1-c83a5639e592-kube-api-access-b6gw4\") pod \"packageserver-d55dfcdfc-nhq52\" (UID: \"6aa7a74a-b3d4-47df-a1b1-c83a5639e592\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.581048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq9lz\" (UniqueName: \"kubernetes.io/projected/4edd042a-f910-49c2-9220-56a6e79b04dc-kube-api-access-hq9lz\") pod \"collect-profiles-29555640-5965f\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.583945 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c68kr"] Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.584740 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.591797 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.596841 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pshhr\" (UniqueName: \"kubernetes.io/projected/0f14f5e4-148e-4282-9243-89c96906048a-kube-api-access-pshhr\") pod \"package-server-manager-789f6589d5-jfb58\" (UID: \"0f14f5e4-148e-4282-9243-89c96906048a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.603570 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.606527 4926 generic.go:334] "Generic (PLEG): container finished" podID="48a0fa25-6b2d-4668-b8e5-824912077f19" containerID="eca16552cabc66a957cf8d61cb7d1632766bc219c67a7d16dae3bf783cc793a0" exitCode=0 Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.606593 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" event={"ID":"48a0fa25-6b2d-4668-b8e5-824912077f19","Type":"ContainerDied","Data":"eca16552cabc66a957cf8d61cb7d1632766bc219c67a7d16dae3bf783cc793a0"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.606618 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" event={"ID":"48a0fa25-6b2d-4668-b8e5-824912077f19","Type":"ContainerStarted","Data":"7a5c5c5c226a6e688b1683c800ce32947a83286c8aff57cfd8b310a771f82f11"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.614768 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" event={"ID":"7ee39302-0316-4481-871e-538ffd31a507","Type":"ContainerStarted","Data":"e024d0e53c7fa71e7e203c8c417d25cff769541c875faeb7d565736258629fed"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.614811 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" event={"ID":"7ee39302-0316-4481-871e-538ffd31a507","Type":"ContainerStarted","Data":"83e3807119b1915ea1314bd29b524a6cebc9ce445f044b1fd3296040625c0c21"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.615673 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.618634 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" event={"ID":"eedd886d-5443-47e1-afbf-5aff90067f3b","Type":"ContainerStarted","Data":"9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.618676 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" event={"ID":"eedd886d-5443-47e1-afbf-5aff90067f3b","Type":"ContainerStarted","Data":"f375d41ba80f860b9235fe8ebb0d083845e27083ae0a2e004437a84b6b1f824e"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.619582 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.625327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrq7x\" (UniqueName: \"kubernetes.io/projected/a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b-kube-api-access-nrq7x\") pod \"machine-config-server-drmq2\" (UID: \"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b\") " pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.628880 4926 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7rmsn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.628933 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" podUID="eedd886d-5443-47e1-afbf-5aff90067f3b" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.629224 4926 patch_prober.go:28] interesting pod/console-operator-58897d9998-s9tvm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.629244 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" podUID="7ee39302-0316-4481-871e-538ffd31a507" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.629465 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.629768 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.129748876 +0000 UTC m=+226.498375209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.640169 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-pd4lm" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.647903 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wndvq" event={"ID":"e379fe1d-7780-4f17-8df8-f74f3dddbc23","Type":"ContainerStarted","Data":"251116173e0184dabcae9f1470dfeba148eb84e6d6fd0d2f14cd2177ac791148"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.647940 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-wndvq" event={"ID":"e379fe1d-7780-4f17-8df8-f74f3dddbc23","Type":"ContainerStarted","Data":"8ecdaf5b859f4687483dedd414756336f848c04f6ad974e42c6606b8128fe624"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.659561 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7"] Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.669321 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl7pp\" (UniqueName: \"kubernetes.io/projected/d68160cf-4e6c-4294-bfdc-4acb74637ecb-kube-api-access-tl7pp\") pod \"auto-csr-approver-29555646-wqpkb\" (UID: \"d68160cf-4e6c-4294-bfdc-4acb74637ecb\") " pod="openshift-infra/auto-csr-approver-29555646-wqpkb" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.670278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-st67x" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.710934 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd"] Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.723704 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" event={"ID":"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e","Type":"ContainerStarted","Data":"25d4891e318db63842a21893f47245b10054234a14851c794f66da64eb9bf8d4"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.723796 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" event={"ID":"5cae3622-73dc-43ee-9e5c-eb6c67e37c1e","Type":"ContainerStarted","Data":"71e522f800a2f7ad15f4527ee0006ccf2a2a361eee826af72fe7e9f03810ff65"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.729641 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" event={"ID":"4230f869-9456-44a1-87b3-342fc8c18ed7","Type":"ContainerStarted","Data":"8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.729675 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" event={"ID":"4230f869-9456-44a1-87b3-342fc8c18ed7","Type":"ContainerStarted","Data":"01692fb622a525e9c42adef3bc2dd9239f99fa47f15b51038978d9cd4978f5eb"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.730286 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.730707 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.731512 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.23150033 +0000 UTC m=+226.600126663 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.738244 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" event={"ID":"5212edbb-b23e-44b8-aab1-77dece6d1415","Type":"ContainerStarted","Data":"028109e363337ff53ceee7c1b43dedf9da3bb3f4229d2026205509d87ba233d4"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.754683 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" event={"ID":"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a","Type":"ContainerStarted","Data":"0db594dcd1a582cafdacd02c1df7aaeaa57401d84254c36a469d20dcf1ceacb5"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.754744 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" event={"ID":"64a2a31c-e6c6-47da-8c6c-2f1ecc2cae0a","Type":"ContainerStarted","Data":"728c89fc3a0575b8e261e5e64c1d58ab08fc1123b8b8e394acde8c8ea8a93621"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.756816 4926 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-svrln container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.757146 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" podUID="4230f869-9456-44a1-87b3-342fc8c18ed7" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.758659 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" event={"ID":"7e94e6e0-b16d-462f-b791-ba20acdcb809","Type":"ContainerStarted","Data":"67c4943e95a1c40606c31e36e0635b0bf125bbcde238649f4a46a977851332cc"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.768056 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" event={"ID":"8c30a7ad-f92b-445a-9201-fe55f247cf41","Type":"ContainerStarted","Data":"1a16a2f2c259d340eac988b4f95f87105acb91d2ac8dd2880085905b12018f2b"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.768092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" event={"ID":"8c30a7ad-f92b-445a-9201-fe55f247cf41","Type":"ContainerStarted","Data":"b55657cb73acd5feec151622d460258d2ed79fb0c98212b3ade5d94e9bbe6b62"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.777331 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" event={"ID":"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca","Type":"ContainerStarted","Data":"7e99758c6494f5e0e5da39dab8d3b4fd2c2d8f5b3e5308d24dc5c55220bef34b"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.777374 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" event={"ID":"a5e9834d-5aeb-4154-ae70-c2b6b07c9eca","Type":"ContainerStarted","Data":"2bd1b4dbef2d5299c5d28d919ce85c08f87b2399a31de3d7c1b45884c90c69cc"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.783663 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" event={"ID":"49e5e304-df7c-434b-8b17-f520e9bb7d52","Type":"ContainerStarted","Data":"6d5e48f8f77cc22923aa8c316c194b01cdb7b43f3f11516201fa35dbdcef9239"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.783729 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" event={"ID":"49e5e304-df7c-434b-8b17-f520e9bb7d52","Type":"ContainerStarted","Data":"3d386874795ceeea475021dc8c00ce587b948725302ec554b6280b4d4f5600da"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.788022 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" event={"ID":"55b6dd11-c219-4b24-90eb-dbc096a67835","Type":"ContainerStarted","Data":"7562a0b82dc71e945dc3015814e8b06ec357f6e7d2119f7931077ec6cb598845"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.788113 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" event={"ID":"55b6dd11-c219-4b24-90eb-dbc096a67835","Type":"ContainerStarted","Data":"13e98918b67aa3a4604fb14453bb42264b92bf6862b1423d05244e57bc41de33"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.813805 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" event={"ID":"4c2f00a8-c3ce-4957-9093-d8c2cce49992","Type":"ContainerStarted","Data":"91cd7543854acab335d02d2b9cdb6610920076420f11b25cf3564733015dfdcd"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.813866 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" event={"ID":"4c2f00a8-c3ce-4957-9093-d8c2cce49992","Type":"ContainerStarted","Data":"46ec498673358f957eb27199406a9e3524f91fc376937de73c5dd4885e364b77"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.823215 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.842068 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.844799 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.344784224 +0000 UTC m=+226.713410547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.864104 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.866777 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.866852 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.867024 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.876044 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.879689 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.882929 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vb9qx" event={"ID":"270031fa-3d83-4edf-bb5d-19ce9e1a693d","Type":"ContainerStarted","Data":"32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b"} Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.913390 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-drmq2" Mar 12 18:06:25 crc kubenswrapper[4926]: I0312 18:06:25.945022 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:25 crc kubenswrapper[4926]: E0312 18:06:25.951008 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.450983951 +0000 UTC m=+226.819610284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.046898 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.048568 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.548545377 +0000 UTC m=+226.917171700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.123996 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.129360 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.133256 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q49l5"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.149593 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.150474 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.650422374 +0000 UTC m=+227.019048707 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.168033 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.169672 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.179679 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.192468 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-bsvxw"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.255628 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.256060 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.756042585 +0000 UTC m=+227.124668918 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.354011 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38004: no serving certificate available for the kubelet" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.358969 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.359710 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.859695951 +0000 UTC m=+227.228322284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.438174 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q2smg" podStartSLOduration=155.438147695 podStartE2EDuration="2m35.438147695s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:26.436893471 +0000 UTC m=+226.805519804" watchObservedRunningTime="2026-03-12 18:06:26.438147695 +0000 UTC m=+226.806774028" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.438291 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38016: no serving certificate available for the kubelet" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.476233 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.476924 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:26.976899904 +0000 UTC m=+227.345526237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.509822 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-hwwdz" podStartSLOduration=155.509802961 podStartE2EDuration="2m35.509802961s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:26.473232342 +0000 UTC m=+226.841858675" watchObservedRunningTime="2026-03-12 18:06:26.509802961 +0000 UTC m=+226.878429294" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.534981 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38648: no serving certificate available for the kubelet" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.589377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.589753 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.089739416 +0000 UTC m=+227.458365749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.609038 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-x7fm2" podStartSLOduration=155.609018943 podStartE2EDuration="2m35.609018943s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:26.607898991 +0000 UTC m=+226.976525324" watchObservedRunningTime="2026-03-12 18:06:26.609018943 +0000 UTC m=+226.977645266" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.643021 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38664: no serving certificate available for the kubelet" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.692062 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.719154 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.219123089 +0000 UTC m=+227.587749422 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.755761 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38668: no serving certificate available for the kubelet" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.818767 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.818853 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.819248 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.819103 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.319091482 +0000 UTC m=+227.687717815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.865274 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38678: no serving certificate available for the kubelet" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.868665 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-wndvq" podStartSLOduration=155.868642801 podStartE2EDuration="2m35.868642801s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:26.810611456 +0000 UTC m=+227.179237809" watchObservedRunningTime="2026-03-12 18:06:26.868642801 +0000 UTC m=+227.237269134" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.876852 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:26 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:26 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:26 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.877035 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.897133 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" podStartSLOduration=155.897118545 podStartE2EDuration="2m35.897118545s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:26.894923734 +0000 UTC m=+227.263550067" watchObservedRunningTime="2026-03-12 18:06:26.897118545 +0000 UTC m=+227.265744878" Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.913087 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.922835 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.922914 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.422895883 +0000 UTC m=+227.791522216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.923572 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:26 crc kubenswrapper[4926]: E0312 18:06:26.924107 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.424085096 +0000 UTC m=+227.792711429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.941364 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9"] Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.949853 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" event={"ID":"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b","Type":"ContainerStarted","Data":"7a6f3ba59fc7e8583a370a6e58f7ad7026a2677d263e98044721d21503f15589"} Mar 12 18:06:26 crc kubenswrapper[4926]: I0312 18:06:26.974662 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-qbxxz" podStartSLOduration=155.974636993 podStartE2EDuration="2m35.974636993s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:26.960205052 +0000 UTC m=+227.328831385" watchObservedRunningTime="2026-03-12 18:06:26.974636993 +0000 UTC m=+227.343263336" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.006704 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-drmq2" event={"ID":"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b","Type":"ContainerStarted","Data":"97bc5d944c909a6163942dd9455d7d8952973ac21f1d08511a5d86cdab39077e"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.016600 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38690: no serving certificate available for the kubelet" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.021293 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" event={"ID":"f54babbf-5b93-4ef7-99d7-0cfe05e921c5","Type":"ContainerStarted","Data":"bf119e4b4a98067cffceb4ed59a629041510e11eadc38d9a74a3cd8f3bfbd2a6"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.024577 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.024895 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.524879323 +0000 UTC m=+227.893505656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.037100 4926 generic.go:334] "Generic (PLEG): container finished" podID="7e94e6e0-b16d-462f-b791-ba20acdcb809" containerID="111f37543ab21017395918e0ccd1df09471d3fe16ac7e1da8c73a6cc9fe06a30" exitCode=0 Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.037288 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" event={"ID":"7e94e6e0-b16d-462f-b791-ba20acdcb809","Type":"ContainerDied","Data":"111f37543ab21017395918e0ccd1df09471d3fe16ac7e1da8c73a6cc9fe06a30"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.089643 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" event={"ID":"b99fa2a7-b47e-433c-b6d5-7c1306a79cde","Type":"ContainerStarted","Data":"46dff74c54b464b95a14e1b9b5858f50f013957acfa111c498373bbabe094b26"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.106290 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-bfn44" podStartSLOduration=156.106265728 podStartE2EDuration="2m36.106265728s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.090070447 +0000 UTC m=+227.458696780" watchObservedRunningTime="2026-03-12 18:06:27.106265728 +0000 UTC m=+227.474892061" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.126609 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.127209 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.627184051 +0000 UTC m=+227.995810384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.144604 4926 generic.go:334] "Generic (PLEG): container finished" podID="5212edbb-b23e-44b8-aab1-77dece6d1415" containerID="0714bb4dcff4dc2428df4aeda6780222f34109de8c6b09d77bc70541c06afb17" exitCode=0 Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.145238 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" event={"ID":"5212edbb-b23e-44b8-aab1-77dece6d1415","Type":"ContainerDied","Data":"0714bb4dcff4dc2428df4aeda6780222f34109de8c6b09d77bc70541c06afb17"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.185947 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-st67x"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.191335 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" event={"ID":"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51","Type":"ContainerStarted","Data":"b9f1222ee65a38125a7545024c47563dcf35849967132a1c5311a2a6a6a7e885"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.205308 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" event={"ID":"a931c621-236d-4a2f-9c96-2c29483f19db","Type":"ContainerStarted","Data":"9de569ecb6b66b1d5452a6d02bfa27e9f6d654e0af62852910db34e88fe9bf9a"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.208212 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-pd4lm"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.225526 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vb9qx" podStartSLOduration=156.225504758 podStartE2EDuration="2m36.225504758s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.213882475 +0000 UTC m=+227.582508818" watchObservedRunningTime="2026-03-12 18:06:27.225504758 +0000 UTC m=+227.594131091" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.225649 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-rxml7"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.235024 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.235704 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.735656381 +0000 UTC m=+228.104282724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.240679 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.252811 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" event={"ID":"439a0925-eff6-4da1-b65c-de0623b9daaa","Type":"ContainerStarted","Data":"70134ce23df9422d4c083805e0f21dd102ee5a93120340676eb68da71536b00c"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.289059 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.291475 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" event={"ID":"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98","Type":"ContainerStarted","Data":"0e10ddf44702564bdb0735ff145d6f01e01df0a99bcf4c2caaa2097751e18b0f"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.300679 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-2wsbc" podStartSLOduration=156.300660051 podStartE2EDuration="2m36.300660051s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.291044613 +0000 UTC m=+227.659670946" watchObservedRunningTime="2026-03-12 18:06:27.300660051 +0000 UTC m=+227.669286394" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.315090 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" event={"ID":"1dc82997-2782-4e9e-a293-956fcb96acde","Type":"ContainerStarted","Data":"806e08739a54e95bdb8abf4febfb3f8c2188846a99469ff5f9e5ae8b483c7d80"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.332150 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" podStartSLOduration=156.332133337 podStartE2EDuration="2m36.332133337s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.328871616 +0000 UTC m=+227.697497949" watchObservedRunningTime="2026-03-12 18:06:27.332133337 +0000 UTC m=+227.700759670" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.337405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.337517 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rx7tc"] Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.337722 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.837711723 +0000 UTC m=+228.206338056 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.339208 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.363784 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-c9xl5"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.379216 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qs58t" podStartSLOduration=156.379198198 podStartE2EDuration="2m36.379198198s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.378468287 +0000 UTC m=+227.747094620" watchObservedRunningTime="2026-03-12 18:06:27.379198198 +0000 UTC m=+227.747824531" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.384946 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38700: no serving certificate available for the kubelet" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.412836 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555646-wqpkb"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.443801 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.443988 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.943957171 +0000 UTC m=+228.312583494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.444258 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.445169 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:27.945158284 +0000 UTC m=+228.313784617 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.456608 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" event={"ID":"fb73641e-b156-470d-be1a-52c11f2efdf6","Type":"ContainerStarted","Data":"ab450526f0818681745b07bb50413cd15f837f3f97782151907eedba15623950"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.460095 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.462426 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.463892 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" event={"ID":"36a0d781-e388-41ae-878e-05a69d81c83e","Type":"ContainerStarted","Data":"eb51896bc46a5a937f2c8797ccaef53c9c5e609a899b6bd2d5834f33c221ee00"} Mar 12 18:06:27 crc kubenswrapper[4926]: W0312 18:06:27.528594 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd68160cf_4e6c_4294_bfdc_4acb74637ecb.slice/crio-71f803a395609858e2670b2b53c7204676afa8a634426e199a4c62b07d977042 WatchSource:0}: Error finding container 71f803a395609858e2670b2b53c7204676afa8a634426e199a4c62b07d977042: Status 404 returned error can't find the container with id 71f803a395609858e2670b2b53c7204676afa8a634426e199a4c62b07d977042 Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.533014 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" podStartSLOduration=156.53299947 podStartE2EDuration="2m36.53299947s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.50640379 +0000 UTC m=+227.875030123" watchObservedRunningTime="2026-03-12 18:06:27.53299947 +0000 UTC m=+227.901625803" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.533131 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" event={"ID":"5d073c88-4608-4594-9feb-f1093455368d","Type":"ContainerStarted","Data":"c6f7764c266bdcce29b8bcda278c7f93167606bdf942f0af18206615b3309291"} Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.541175 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58"] Mar 12 18:06:27 crc kubenswrapper[4926]: W0312 18:06:27.540779 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8acca805_2bf8_4dcb_a036_c4732084210e.slice/crio-8b631f999700f2f51aa43e2ed2f5b2edb1d58820b4c614d63e8892b10d5e4413 WatchSource:0}: Error finding container 8b631f999700f2f51aa43e2ed2f5b2edb1d58820b4c614d63e8892b10d5e4413: Status 404 returned error can't find the container with id 8b631f999700f2f51aa43e2ed2f5b2edb1d58820b4c614d63e8892b10d5e4413 Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.545365 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.545953 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.04593483 +0000 UTC m=+228.414561163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.553961 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.553971 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.554410 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.602743 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.651058 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.659216 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6hxxs" podStartSLOduration=156.659192684 podStartE2EDuration="2m36.659192684s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.652290001 +0000 UTC m=+228.020916334" watchObservedRunningTime="2026-03-12 18:06:27.659192684 +0000 UTC m=+228.027819017" Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.660809 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.160793038 +0000 UTC m=+228.529419371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.707586 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q49l5"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.738717 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln"] Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.752774 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.753130 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.253113469 +0000 UTC m=+228.621739802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.798271 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-9zg78" podStartSLOduration=156.798214995 podStartE2EDuration="2m36.798214995s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:27.79264997 +0000 UTC m=+228.161276323" watchObservedRunningTime="2026-03-12 18:06:27.798214995 +0000 UTC m=+228.166841328" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.855374 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.856257 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.356242561 +0000 UTC m=+228.724868894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.868749 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:27 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:27 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:27 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.868835 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.956350 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.956526 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.456490552 +0000 UTC m=+228.825116885 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:27 crc kubenswrapper[4926]: I0312 18:06:27.957224 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:27 crc kubenswrapper[4926]: E0312 18:06:27.957575 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.457563421 +0000 UTC m=+228.826189744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.058608 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.058968 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.558925984 +0000 UTC m=+228.927552317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.082323 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38704: no serving certificate available for the kubelet" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.163154 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.163788 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.663772284 +0000 UTC m=+229.032398607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.164338 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-s9tvm" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.264660 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.264832 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.764801777 +0000 UTC m=+229.133428130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.264929 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.265202 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.765191917 +0000 UTC m=+229.133818250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.366345 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.366884 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.866868798 +0000 UTC m=+229.235495121 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.467951 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.468221 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:28.96821031 +0000 UTC m=+229.336836643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.566665 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-st67x" event={"ID":"06394469-defd-4710-90a2-b6c395c00d4f","Type":"ContainerStarted","Data":"0b346b108d62eb5da49953967efe8d2aa508da3a0686aee8e0074faddf6a0deb"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.566963 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-st67x" event={"ID":"06394469-defd-4710-90a2-b6c395c00d4f","Type":"ContainerStarted","Data":"c3c4ba69a1afa467197ea96769c8eddc625df688c9e556cde4a4df95630be0c7"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.567910 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-st67x" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.570776 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.571062 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.071048764 +0000 UTC m=+229.439675097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.571130 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-st67x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.571154 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-st67x" podUID="06394469-defd-4710-90a2-b6c395c00d4f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.582719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" event={"ID":"48a0fa25-6b2d-4668-b8e5-824912077f19","Type":"ContainerStarted","Data":"902017b816fa06868e32faf2a15deecb13941d5338ff7a5ccdca3d1cef972aaa"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.584429 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" event={"ID":"6aa7a74a-b3d4-47df-a1b1-c83a5639e592","Type":"ContainerStarted","Data":"d02ba0167b9d3dcca2129478af6fb364197237d858ab6b10ee3c1903d568411b"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.585801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" event={"ID":"439a0925-eff6-4da1-b65c-de0623b9daaa","Type":"ContainerStarted","Data":"706bb3ae8d83a05d06d62a552dbf2b2212addee1d94ab04aa29717c7aa168687"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.585821 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" event={"ID":"439a0925-eff6-4da1-b65c-de0623b9daaa","Type":"ContainerStarted","Data":"02ddaa7c3b1f7c1e19e74cc27305fa6152d7ea38439bde76362b6a569f7807ab"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.587808 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" event={"ID":"f54babbf-5b93-4ef7-99d7-0cfe05e921c5","Type":"ContainerStarted","Data":"063c70ab5812212bb3956af2f3aadc7cbf5842847da96dcb5a138072de4412d4"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.591451 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-st67x" podStartSLOduration=157.591422731 podStartE2EDuration="2m37.591422731s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.588734626 +0000 UTC m=+228.957360969" watchObservedRunningTime="2026-03-12 18:06:28.591422731 +0000 UTC m=+228.960049064" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.616259 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" event={"ID":"7e94e6e0-b16d-462f-b791-ba20acdcb809","Type":"ContainerStarted","Data":"7b6f93f7edb6492e4695d79094d85d1f0d7cae2a51625b9c7258b1a05824775b"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.619092 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-snpzd" podStartSLOduration=157.619069291 podStartE2EDuration="2m37.619069291s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.617182249 +0000 UTC m=+228.985808582" watchObservedRunningTime="2026-03-12 18:06:28.619069291 +0000 UTC m=+228.987695634" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.635143 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" event={"ID":"5212edbb-b23e-44b8-aab1-77dece6d1415","Type":"ContainerStarted","Data":"3683e6d87ebfe0ba92a5a25a982d7182088dee633daa02366e090901ac184a34"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.635695 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.637364 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" event={"ID":"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba","Type":"ContainerStarted","Data":"c894dbfb8f955a86829df2ba4efcf318943892ea413eff807fd583ce708825e5"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.637383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" event={"ID":"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba","Type":"ContainerStarted","Data":"03009e64aa09225c59aea596418228aebd555b8d4b2b3d4aac4ce13dbcb9ad9f"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.637392 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" event={"ID":"2e2b819b-3df8-4d51-b6d0-b30e33f1ceba","Type":"ContainerStarted","Data":"17b32e16b34462b3c5ef720559a7456983ee51be0128c62b7047131940a33bc0"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.639240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" event={"ID":"0f14f5e4-148e-4282-9243-89c96906048a","Type":"ContainerStarted","Data":"50d50f4e8d969e7d550102332232f39ac77c2b02640011b745a39d32eeece517"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.639261 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" event={"ID":"0f14f5e4-148e-4282-9243-89c96906048a","Type":"ContainerStarted","Data":"b65b9aeced60e6b54a95ee1f5501d459c45e422db6fa17a55783725f478f175f"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.639889 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.643841 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" event={"ID":"00918e77-d6a1-4d1b-9986-37f3e648b322","Type":"ContainerStarted","Data":"a38ba83fac41229554dec49b5d2efd44825a6028f22c7a88f041d1ef43dea75f"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.643866 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" event={"ID":"00918e77-d6a1-4d1b-9986-37f3e648b322","Type":"ContainerStarted","Data":"efcb66960883ede90dc195e440865f7e82c0b5f56d998bbc51bade6dcf61a851"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.651863 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" event={"ID":"8acca805-2bf8-4dcb-a036-c4732084210e","Type":"ContainerStarted","Data":"8b631f999700f2f51aa43e2ed2f5b2edb1d58820b4c614d63e8892b10d5e4413"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.657466 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" event={"ID":"a931c621-236d-4a2f-9c96-2c29483f19db","Type":"ContainerStarted","Data":"df7538fc54525f099380b778ea1af02a79e3048b1fbc5d5925880e93ce1305aa"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.662045 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sbn5h" podStartSLOduration=157.662029177 podStartE2EDuration="2m37.662029177s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.642401681 +0000 UTC m=+229.011028014" watchObservedRunningTime="2026-03-12 18:06:28.662029177 +0000 UTC m=+229.030655510" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.672584 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.674073 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" event={"ID":"fb73641e-b156-470d-be1a-52c11f2efdf6","Type":"ContainerStarted","Data":"fa0d8501714fa5a206cbf7622a0a89cc8e4757cce153b86a3505d49984e2488f"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.674176 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" event={"ID":"fb73641e-b156-470d-be1a-52c11f2efdf6","Type":"ContainerStarted","Data":"3fcc7bbe3084bb0081beb292a638fe49741f558a7fe0817779cbbfcad3ca40dd"} Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.675706 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.175695228 +0000 UTC m=+229.544321561 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.677066 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" event={"ID":"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51","Type":"ContainerStarted","Data":"9e9ac9fc823dfa71afcfab8c0327acbbbd68667ed466479a9ab182c4d3a10028"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.681329 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" podStartSLOduration=157.681315904 podStartE2EDuration="2m37.681315904s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.665037831 +0000 UTC m=+229.033664164" watchObservedRunningTime="2026-03-12 18:06:28.681315904 +0000 UTC m=+229.049942237" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.681735 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" podStartSLOduration=157.681730126 podStartE2EDuration="2m37.681730126s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.680345967 +0000 UTC m=+229.048972310" watchObservedRunningTime="2026-03-12 18:06:28.681730126 +0000 UTC m=+229.050356459" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.683542 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" event={"ID":"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b","Type":"ContainerStarted","Data":"c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.684335 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.686298 4926 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-c68kr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.686335 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.692189 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rxml7" event={"ID":"4c9eaabb-c0c2-4f71-82eb-7494d0a37075","Type":"ContainerStarted","Data":"cf0dcaa07c4b2300045723a181b0c00320c05ebaa5b17c3d50437dd6250eb5db"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.692226 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rxml7" event={"ID":"4c9eaabb-c0c2-4f71-82eb-7494d0a37075","Type":"ContainerStarted","Data":"ce252b86e505beb87d6ebc7320313de29a639b0e114586d0a2f5174a585af05e"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.693249 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" event={"ID":"5d073c88-4608-4594-9feb-f1093455368d","Type":"ContainerStarted","Data":"d77934aa0f03829fb16e2ee2875c4c7bb8bd67808e9108f3762f5211b24f4d70"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.694766 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" event={"ID":"d68160cf-4e6c-4294-bfdc-4acb74637ecb","Type":"ContainerStarted","Data":"71f803a395609858e2670b2b53c7204676afa8a634426e199a4c62b07d977042"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.696059 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" event={"ID":"416787ca-b62c-4fd5-84c5-57be59317faa","Type":"ContainerStarted","Data":"bc0c7e300e9f1ca23244b9b1e4f774aed1405acdaf5be46e0f9cca806457d750"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.696079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" event={"ID":"416787ca-b62c-4fd5-84c5-57be59317faa","Type":"ContainerStarted","Data":"4ae398fd482d6448935b238177f877b46b8c26bd31ee1f2f7798f802d6d585f7"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.709048 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pd4lm" event={"ID":"c574438a-86a5-4a9f-aff3-47bf920cbabd","Type":"ContainerStarted","Data":"499c10c00a7a67b5c7de38ccfaf74c7acd1f77f0a75dbbc45898fba6fc3872b5"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.709086 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-pd4lm" event={"ID":"c574438a-86a5-4a9f-aff3-47bf920cbabd","Type":"ContainerStarted","Data":"d6fb422db11ac7fcfc5af5930643d0c8e176ffffc4dd20b87f46927d4c33fcb1"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.713221 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" event={"ID":"36a0d781-e388-41ae-878e-05a69d81c83e","Type":"ContainerStarted","Data":"d7b30f818d802e3531db55adca1135dc83e00abca124fbeb2915f0f83c88fa84"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.717472 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" podStartSLOduration=157.71742457 podStartE2EDuration="2m37.71742457s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.717042449 +0000 UTC m=+229.085668792" watchObservedRunningTime="2026-03-12 18:06:28.71742457 +0000 UTC m=+229.086050903" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.728952 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" event={"ID":"721fd54d-55f5-4477-a7b8-283a700c8c47","Type":"ContainerStarted","Data":"c069c12da1b73cd3855cc706b05648787a3c2f323859e6419a8cdbb2cc4085fa"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.728985 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" event={"ID":"721fd54d-55f5-4477-a7b8-283a700c8c47","Type":"ContainerStarted","Data":"f632cc51bbf1817e9bf02cf814a40e910839cfb355f7afa5f22369074b6a5e51"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.736364 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-bmp6d" podStartSLOduration=157.736348856 podStartE2EDuration="2m37.736348856s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.734874866 +0000 UTC m=+229.103501209" watchObservedRunningTime="2026-03-12 18:06:28.736348856 +0000 UTC m=+229.104975189" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.743382 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" event={"ID":"f4867106-501e-4408-af0b-7790cfc45a24","Type":"ContainerStarted","Data":"893f7549880003c3bed9600e5e303162fde1f9a3b368133025d5441193467a5c"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.743422 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" event={"ID":"f4867106-501e-4408-af0b-7790cfc45a24","Type":"ContainerStarted","Data":"fc6fe3ee88adf1a106f670dc0abbabe0e6215202e5d0a3834c479f5e9ad5dd1a"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.746362 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" event={"ID":"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98","Type":"ContainerStarted","Data":"239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.746895 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.748202 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" event={"ID":"f6ad9439-d12c-4987-840a-002975ba1498","Type":"ContainerStarted","Data":"c20e9eb36d7549aa43e55a83b069e75fd8e5138b013ae4e7628f337a9ab5601a"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.748230 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" event={"ID":"f6ad9439-d12c-4987-840a-002975ba1498","Type":"ContainerStarted","Data":"423ab1f7663178dcad7815397c58f06889c819795f4bfa80e43f7eec42b21289"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.748840 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.750788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" event={"ID":"b99fa2a7-b47e-433c-b6d5-7c1306a79cde","Type":"ContainerStarted","Data":"7c2088451807c045409985d9ad0726d5bdce2c87ca611606a3f1d3b906f1416c"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.751461 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.755090 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" event={"ID":"4edd042a-f910-49c2-9220-56a6e79b04dc","Type":"ContainerStarted","Data":"2317c39f2953accfac56fb1efd908ec6062a7797866f162790fc901a5c9318d9"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.755122 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" event={"ID":"4edd042a-f910-49c2-9220-56a6e79b04dc","Type":"ContainerStarted","Data":"bf6d0c0b52a287d964339cdd9c78d27ee7b19ab7a4e80185b1b7ae31464e1f68"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.758183 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-drmq2" event={"ID":"a9fa6603-7fa6-4f96-9a3d-b6d3b0a0cd5b","Type":"ContainerStarted","Data":"1037edb6d6d02c15c85c19b1ac0f8d86103e850dccee7e65c579bbb97398254b"} Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.764248 4926 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rdjk9 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.764306 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" podUID="b99fa2a7-b47e-433c-b6d5-7c1306a79cde" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.764602 4926 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-x4kkc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.764619 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" podUID="f6ad9439-d12c-4987-840a-002975ba1498" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.764662 4926 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-q49l5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.764676 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" podUID="ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.777579 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.778395 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.278374756 +0000 UTC m=+229.647001099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.798814 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-sgl9z" podStartSLOduration=157.798796115 podStartE2EDuration="2m37.798796115s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.76879537 +0000 UTC m=+229.137421703" watchObservedRunningTime="2026-03-12 18:06:28.798796115 +0000 UTC m=+229.167422438" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.812862 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" podStartSLOduration=157.812842426 podStartE2EDuration="2m37.812842426s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.799623999 +0000 UTC m=+229.168250332" watchObservedRunningTime="2026-03-12 18:06:28.812842426 +0000 UTC m=+229.181468749" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.829090 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-b9s9h" podStartSLOduration=157.829067688 podStartE2EDuration="2m37.829067688s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.816327724 +0000 UTC m=+229.184954057" watchObservedRunningTime="2026-03-12 18:06:28.829067688 +0000 UTC m=+229.197694021" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.851632 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-z4g7g" podStartSLOduration=157.851607876 podStartE2EDuration="2m37.851607876s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.85103896 +0000 UTC m=+229.219665293" watchObservedRunningTime="2026-03-12 18:06:28.851607876 +0000 UTC m=+229.220234209" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.885805 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rx7tc" podStartSLOduration=157.885789118 podStartE2EDuration="2m37.885789118s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.885355805 +0000 UTC m=+229.253982148" watchObservedRunningTime="2026-03-12 18:06:28.885789118 +0000 UTC m=+229.254415451" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.886474 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.886861 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.386845726 +0000 UTC m=+229.755472059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.890873 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:28 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:28 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:28 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.891276 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.910425 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sj8r7" podStartSLOduration=157.910400593 podStartE2EDuration="2m37.910400593s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.906945036 +0000 UTC m=+229.275571389" watchObservedRunningTime="2026-03-12 18:06:28.910400593 +0000 UTC m=+229.279026926" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.931886 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-pd4lm" podStartSLOduration=6.93186608 podStartE2EDuration="6.93186608s" podCreationTimestamp="2026-03-12 18:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.928475116 +0000 UTC m=+229.297101469" watchObservedRunningTime="2026-03-12 18:06:28.93186608 +0000 UTC m=+229.300492413" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.971843 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-h5mf8" podStartSLOduration=157.971827793 podStartE2EDuration="2m37.971827793s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.958746259 +0000 UTC m=+229.327372592" watchObservedRunningTime="2026-03-12 18:06:28.971827793 +0000 UTC m=+229.340454126" Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.988370 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.988538 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.488515528 +0000 UTC m=+229.857141871 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.988718 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:28 crc kubenswrapper[4926]: E0312 18:06:28.989358 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.489330361 +0000 UTC m=+229.857956694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:28 crc kubenswrapper[4926]: I0312 18:06:28.995756 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-drmq2" podStartSLOduration=6.995727009 podStartE2EDuration="6.995727009s" podCreationTimestamp="2026-03-12 18:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:28.993149056 +0000 UTC m=+229.361775409" watchObservedRunningTime="2026-03-12 18:06:28.995727009 +0000 UTC m=+229.364353352" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.035216 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" podStartSLOduration=158.035193317 podStartE2EDuration="2m38.035193317s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.015569351 +0000 UTC m=+229.384195714" watchObservedRunningTime="2026-03-12 18:06:29.035193317 +0000 UTC m=+229.403819650" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.063096 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" podStartSLOduration=158.063067203 podStartE2EDuration="2m38.063067203s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.06185155 +0000 UTC m=+229.430477903" watchObservedRunningTime="2026-03-12 18:06:29.063067203 +0000 UTC m=+229.431693536" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.063671 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-xxzt9" podStartSLOduration=158.06366691 podStartE2EDuration="2m38.06366691s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.038609602 +0000 UTC m=+229.407235945" watchObservedRunningTime="2026-03-12 18:06:29.06366691 +0000 UTC m=+229.432293243" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.082417 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" podStartSLOduration=158.082163645 podStartE2EDuration="2m38.082163645s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.080736685 +0000 UTC m=+229.449363018" watchObservedRunningTime="2026-03-12 18:06:29.082163645 +0000 UTC m=+229.450789968" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.090009 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.090416 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.590402185 +0000 UTC m=+229.959028518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.099323 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" podStartSLOduration=158.099305333 podStartE2EDuration="2m38.099305333s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.094354654 +0000 UTC m=+229.462980997" watchObservedRunningTime="2026-03-12 18:06:29.099305333 +0000 UTC m=+229.467931666" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.191615 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.191966 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.691955302 +0000 UTC m=+230.060581635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.215008 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.215062 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.217056 4926 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-9bzkt container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.217104 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" podUID="7e94e6e0-b16d-462f-b791-ba20acdcb809" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.6:8443/livez\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.292209 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.292512 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.792498252 +0000 UTC m=+230.161124585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.393643 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.394006 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.893994138 +0000 UTC m=+230.262620471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.410631 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38706: no serving certificate available for the kubelet" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.494661 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.494905 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.994874236 +0000 UTC m=+230.363500569 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.495122 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.495646 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:29.995618027 +0000 UTC m=+230.364244360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.596835 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.597486 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.097424393 +0000 UTC m=+230.466050726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.597690 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.598359 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.098346908 +0000 UTC m=+230.466973391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.699700 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.699923 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.199890795 +0000 UTC m=+230.568517128 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.700195 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.700617 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.200604865 +0000 UTC m=+230.569231198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.766174 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" event={"ID":"00918e77-d6a1-4d1b-9986-37f3e648b322","Type":"ContainerStarted","Data":"3e51472c7bd12f10625affaa4c1e75ad538e1e3179fc73e200eb25164934a4ab"} Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.770017 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" event={"ID":"8acca805-2bf8-4dcb-a036-c4732084210e","Type":"ContainerStarted","Data":"aaf60efe6696e6166729de2b984c3370705048c3fbc4edc6148242085d71fdf0"} Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.772774 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" event={"ID":"48a0fa25-6b2d-4668-b8e5-824912077f19","Type":"ContainerStarted","Data":"fc95ef35d9461ed11dbc47eae66311fbb2ff949ccc9eb5153beb2fcf366b41fe"} Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.778897 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" event={"ID":"f68a2b5d-036a-43ee-a9f2-dd94d9f51d51","Type":"ContainerStarted","Data":"de5090f0d7a82ccf34bb8bd7e922d4f50aec2f776b8c7b6207b9ff51c590847a"} Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.781401 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" event={"ID":"6aa7a74a-b3d4-47df-a1b1-c83a5639e592","Type":"ContainerStarted","Data":"15607ad279c98c97717385e271a19c69b978cd455519c54ef15b0977d9f167cf"} Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.781700 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.783536 4926 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nhq52 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.783594 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" podUID="6aa7a74a-b3d4-47df-a1b1-c83a5639e592" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.788919 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-rxml7" event={"ID":"4c9eaabb-c0c2-4f71-82eb-7494d0a37075","Type":"ContainerStarted","Data":"e00dd505fdeef5237730f913bf86e06154c5951c844ff5a324b340d0199cacc6"} Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.789029 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.791733 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dnk9m" podStartSLOduration=158.791717122 podStartE2EDuration="2m38.791717122s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.791190227 +0000 UTC m=+230.159816580" watchObservedRunningTime="2026-03-12 18:06:29.791717122 +0000 UTC m=+230.160343455" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.800228 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" event={"ID":"0f14f5e4-148e-4282-9243-89c96906048a","Type":"ContainerStarted","Data":"651b0bd421c4418ad38b69f4049bd85d537ba3d6a63234f7cb3bdda9af139e81"} Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.801149 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-st67x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.801204 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-st67x" podUID="06394469-defd-4710-90a2-b6c395c00d4f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.801149 4926 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-x4kkc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.801284 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" podUID="f6ad9439-d12c-4987-840a-002975ba1498" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.801976 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" podUID="ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" containerName="controller-manager" containerID="cri-o://239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7" gracePeriod=30 Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.802760 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" podUID="4230f869-9456-44a1-87b3-342fc8c18ed7" containerName="route-controller-manager" containerID="cri-o://8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d" gracePeriod=30 Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.804050 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.804215 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.30419875 +0000 UTC m=+230.672825083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.804515 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.806062 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.306050861 +0000 UTC m=+230.674677194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.815856 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.817716 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-rxml7" podStartSLOduration=7.817701056 podStartE2EDuration="7.817701056s" podCreationTimestamp="2026-03-12 18:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.816525572 +0000 UTC m=+230.185151935" watchObservedRunningTime="2026-03-12 18:06:29.817701056 +0000 UTC m=+230.186327389" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.865546 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:29 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:29 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:29 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.865606 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.878903 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rdjk9" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.908348 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.909814 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.40978967 +0000 UTC m=+230.778415993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.909871 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:29 crc kubenswrapper[4926]: E0312 18:06:29.910963 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.410943532 +0000 UTC m=+230.779569865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.953961 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:06:29 crc kubenswrapper[4926]: I0312 18:06:29.959907 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" podStartSLOduration=158.959885384 podStartE2EDuration="2m38.959885384s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:29.954395322 +0000 UTC m=+230.323021655" watchObservedRunningTime="2026-03-12 18:06:29.959885384 +0000 UTC m=+230.328511717" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.030148 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.030760 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.530738738 +0000 UTC m=+230.899365071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.088001 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-bsvxw" podStartSLOduration=159.087981141 podStartE2EDuration="2m39.087981141s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:30.086519481 +0000 UTC m=+230.455145824" watchObservedRunningTime="2026-03-12 18:06:30.087981141 +0000 UTC m=+230.456607464" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.089208 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" podStartSLOduration=159.089202055 podStartE2EDuration="2m39.089202055s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:30.045135639 +0000 UTC m=+230.413761962" watchObservedRunningTime="2026-03-12 18:06:30.089202055 +0000 UTC m=+230.457828388" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.132942 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.133409 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.633385835 +0000 UTC m=+231.002012328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.234255 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.234498 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.73446801 +0000 UTC m=+231.103094363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.234720 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.235112 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.735098658 +0000 UTC m=+231.103724991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.335841 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.336430 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.836417239 +0000 UTC m=+231.205043562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.438307 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.438783 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:30.938758338 +0000 UTC m=+231.307384681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.540001 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.540414 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.040399029 +0000 UTC m=+231.409025362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.582810 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.604028 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.640757 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-proxy-ca-bundles\") pod \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.640911 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-config\") pod \"4230f869-9456-44a1-87b3-342fc8c18ed7\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.641022 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlg4x\" (UniqueName: \"kubernetes.io/projected/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-kube-api-access-nlg4x\") pod \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.641107 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-client-ca\") pod \"4230f869-9456-44a1-87b3-342fc8c18ed7\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.641196 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-client-ca\") pod \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.641276 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-config\") pod \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.641366 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4230f869-9456-44a1-87b3-342fc8c18ed7-serving-cert\") pod \"4230f869-9456-44a1-87b3-342fc8c18ed7\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.641755 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-serving-cert\") pod \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\" (UID: \"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.641862 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28x4l\" (UniqueName: \"kubernetes.io/projected/4230f869-9456-44a1-87b3-342fc8c18ed7-kube-api-access-28x4l\") pod \"4230f869-9456-44a1-87b3-342fc8c18ed7\" (UID: \"4230f869-9456-44a1-87b3-342fc8c18ed7\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.642112 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.642402 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.142391028 +0000 UTC m=+231.511017361 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.643335 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" (UID: "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.643645 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-config" (OuterVolumeSpecName: "config") pod "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" (UID: "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.644308 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-config" (OuterVolumeSpecName: "config") pod "4230f869-9456-44a1-87b3-342fc8c18ed7" (UID: "4230f869-9456-44a1-87b3-342fc8c18ed7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.645393 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-client-ca" (OuterVolumeSpecName: "client-ca") pod "4230f869-9456-44a1-87b3-342fc8c18ed7" (UID: "4230f869-9456-44a1-87b3-342fc8c18ed7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.653394 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-747cdc9dc9-hpk65"] Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.658036 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4230f869-9456-44a1-87b3-342fc8c18ed7" containerName="route-controller-manager" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.658075 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4230f869-9456-44a1-87b3-342fc8c18ed7" containerName="route-controller-manager" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.658094 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" containerName="controller-manager" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.658104 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" containerName="controller-manager" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.658233 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" containerName="controller-manager" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.658256 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4230f869-9456-44a1-87b3-342fc8c18ed7" containerName="route-controller-manager" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.658649 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-client-ca" (OuterVolumeSpecName: "client-ca") pod "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" (UID: "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.658755 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.663676 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4230f869-9456-44a1-87b3-342fc8c18ed7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4230f869-9456-44a1-87b3-342fc8c18ed7" (UID: "4230f869-9456-44a1-87b3-342fc8c18ed7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.664891 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" (UID: "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.665060 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-kube-api-access-nlg4x" (OuterVolumeSpecName: "kube-api-access-nlg4x") pod "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" (UID: "ab2da4b6-fc75-4f97-b27c-3f687ddf9d98"). InnerVolumeSpecName "kube-api-access-nlg4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.699766 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4230f869-9456-44a1-87b3-342fc8c18ed7-kube-api-access-28x4l" (OuterVolumeSpecName: "kube-api-access-28x4l") pod "4230f869-9456-44a1-87b3-342fc8c18ed7" (UID: "4230f869-9456-44a1-87b3-342fc8c18ed7"). InnerVolumeSpecName "kube-api-access-28x4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.707394 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-747cdc9dc9-hpk65"] Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.743891 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744075 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-client-ca\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744103 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-serving-cert\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744132 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-proxy-ca-bundles\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744164 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-config\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hbtx\" (UniqueName: \"kubernetes.io/projected/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-kube-api-access-5hbtx\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744248 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlg4x\" (UniqueName: \"kubernetes.io/projected/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-kube-api-access-nlg4x\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744259 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744269 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744277 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744285 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4230f869-9456-44a1-87b3-342fc8c18ed7-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744295 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744303 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28x4l\" (UniqueName: \"kubernetes.io/projected/4230f869-9456-44a1-87b3-342fc8c18ed7-kube-api-access-28x4l\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744311 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.744319 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4230f869-9456-44a1-87b3-342fc8c18ed7-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.744391 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.244375738 +0000 UTC m=+231.613002061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.846420 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-proxy-ca-bundles\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.846510 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-config\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.846545 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hbtx\" (UniqueName: \"kubernetes.io/projected/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-kube-api-access-5hbtx\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.846577 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.847085 4926 generic.go:334] "Generic (PLEG): container finished" podID="4edd042a-f910-49c2-9220-56a6e79b04dc" containerID="2317c39f2953accfac56fb1efd908ec6062a7797866f162790fc901a5c9318d9" exitCode=0 Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.847148 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" event={"ID":"4edd042a-f910-49c2-9220-56a6e79b04dc","Type":"ContainerDied","Data":"2317c39f2953accfac56fb1efd908ec6062a7797866f162790fc901a5c9318d9"} Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.849064 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-proxy-ca-bundles\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.849136 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-config\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.849191 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-client-ca\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.849225 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-serving-cert\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.849429 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.349413322 +0000 UTC m=+231.718039745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.850338 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-client-ca\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.859588 4926 generic.go:334] "Generic (PLEG): container finished" podID="4230f869-9456-44a1-87b3-342fc8c18ed7" containerID="8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d" exitCode=0 Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.859652 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" event={"ID":"4230f869-9456-44a1-87b3-342fc8c18ed7","Type":"ContainerDied","Data":"8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d"} Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.859679 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" event={"ID":"4230f869-9456-44a1-87b3-342fc8c18ed7","Type":"ContainerDied","Data":"01692fb622a525e9c42adef3bc2dd9239f99fa47f15b51038978d9cd4978f5eb"} Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.859699 4926 scope.go:117] "RemoveContainer" containerID="8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.859856 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.866879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-serving-cert\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.870070 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hbtx\" (UniqueName: \"kubernetes.io/projected/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-kube-api-access-5hbtx\") pod \"controller-manager-747cdc9dc9-hpk65\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.870275 4926 generic.go:334] "Generic (PLEG): container finished" podID="ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" containerID="239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7" exitCode=0 Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.870834 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.873815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" event={"ID":"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98","Type":"ContainerDied","Data":"239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7"} Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.873894 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-q49l5" event={"ID":"ab2da4b6-fc75-4f97-b27c-3f687ddf9d98","Type":"ContainerDied","Data":"0e10ddf44702564bdb0735ff145d6f01e01df0a99bcf4c2caaa2097751e18b0f"} Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.876332 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-st67x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.876370 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-st67x" podUID="06394469-defd-4710-90a2-b6c395c00d4f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.877268 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:30 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:30 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:30 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.877313 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.896770 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-x4kkc" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.912096 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-h95sl" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.942651 4926 scope.go:117] "RemoveContainer" containerID="8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d" Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.943454 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d\": container with ID starting with 8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d not found: ID does not exist" containerID="8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.943488 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d"} err="failed to get container status \"8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d\": rpc error: code = NotFound desc = could not find container \"8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d\": container with ID starting with 8dee9c013e3c6386c5835b2c2d55dd9c7c4c750854cac3e4491b0748bba2145d not found: ID does not exist" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.943511 4926 scope.go:117] "RemoveContainer" containerID="239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.943870 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nhq52" Mar 12 18:06:30 crc kubenswrapper[4926]: I0312 18:06:30.951062 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:30 crc kubenswrapper[4926]: E0312 18:06:30.952272 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.452245116 +0000 UTC m=+231.820871449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.006747 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.011318 4926 scope.go:117] "RemoveContainer" containerID="239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.012685 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7\": container with ID starting with 239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7 not found: ID does not exist" containerID="239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.012719 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7"} err="failed to get container status \"239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7\": rpc error: code = NotFound desc = could not find container \"239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7\": container with ID starting with 239dd029acb9c93acb5f73faccb1f785db98cc3574d192e8eef48910ed524ac7 not found: ID does not exist" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.018483 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-svrln"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.026243 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q49l5"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.031868 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.034687 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-q49l5"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.046655 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s4zf5"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.047611 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.055954 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.061818 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s4zf5"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.070980 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.071358 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.571342442 +0000 UTC m=+231.939968775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.149079 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-565fl"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.150062 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.158596 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.173896 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.174169 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-utilities\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.174242 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7tf\" (UniqueName: \"kubernetes.io/projected/b2a609cd-c298-4356-9ddf-a7f125b52938-kube-api-access-wf7tf\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.174284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-catalog-content\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.174392 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.674375661 +0000 UTC m=+232.043001994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.177883 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-565fl"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.282989 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-catalog-content\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.283032 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7tf\" (UniqueName: \"kubernetes.io/projected/b2a609cd-c298-4356-9ddf-a7f125b52938-kube-api-access-wf7tf\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.283054 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.283085 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wxfn\" (UniqueName: \"kubernetes.io/projected/b5fe4032-6a1e-4c27-9471-fa53e044826e-kube-api-access-2wxfn\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.283106 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-catalog-content\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.283158 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-utilities\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.283207 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-utilities\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.283703 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.783692375 +0000 UTC m=+232.152318698 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.284155 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-catalog-content\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.284354 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-utilities\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.335947 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7tf\" (UniqueName: \"kubernetes.io/projected/b2a609cd-c298-4356-9ddf-a7f125b52938-kube-api-access-wf7tf\") pod \"community-operators-s4zf5\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.350866 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t4sbh"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.351865 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.367788 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t4sbh"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.385004 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.385199 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-utilities\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.385229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-catalog-content\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.385259 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wxfn\" (UniqueName: \"kubernetes.io/projected/b5fe4032-6a1e-4c27-9471-fa53e044826e-kube-api-access-2wxfn\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.386208 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.886162238 +0000 UTC m=+232.254788571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.386327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-utilities\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.386334 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-catalog-content\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.414234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wxfn\" (UniqueName: \"kubernetes.io/projected/b5fe4032-6a1e-4c27-9471-fa53e044826e-kube-api-access-2wxfn\") pod \"certified-operators-565fl\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.416246 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.478360 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-747cdc9dc9-hpk65"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.486528 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnjm4\" (UniqueName: \"kubernetes.io/projected/150781c8-5ae3-42a6-b351-2388dfe84167-kube-api-access-tnjm4\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.486590 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-utilities\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.486612 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-catalog-content\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.486640 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.486885 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:31.986872853 +0000 UTC m=+232.355499186 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.493339 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.539810 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zpnq8"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.540877 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.557411 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpnq8"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.591804 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.592505 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnjm4\" (UniqueName: \"kubernetes.io/projected/150781c8-5ae3-42a6-b351-2388dfe84167-kube-api-access-tnjm4\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.593109 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-utilities\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.593186 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-catalog-content\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.597542 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.097518603 +0000 UTC m=+232.466144936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.598031 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-utilities\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.599089 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-catalog-content\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.617408 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnjm4\" (UniqueName: \"kubernetes.io/projected/150781c8-5ae3-42a6-b351-2388dfe84167-kube-api-access-tnjm4\") pod \"community-operators-t4sbh\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.685731 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.698083 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-utilities\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.698309 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.698492 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsdsv\" (UniqueName: \"kubernetes.io/projected/743a7318-33d0-4a59-93bd-7c6899554e5e-kube-api-access-fsdsv\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.698713 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-catalog-content\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.698850 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.198831644 +0000 UTC m=+232.567458057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.750483 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s4zf5"] Mar 12 18:06:31 crc kubenswrapper[4926]: W0312 18:06:31.799118 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2a609cd_c298_4356_9ddf_a7f125b52938.slice/crio-6da98836531143c0f0c9e5b0f54f53b5a8c52f32a98b1d55d2cd5e57c0ec7c9e WatchSource:0}: Error finding container 6da98836531143c0f0c9e5b0f54f53b5a8c52f32a98b1d55d2cd5e57c0ec7c9e: Status 404 returned error can't find the container with id 6da98836531143c0f0c9e5b0f54f53b5a8c52f32a98b1d55d2cd5e57c0ec7c9e Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.807796 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.808100 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-catalog-content\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.808235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-utilities\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.808318 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsdsv\" (UniqueName: \"kubernetes.io/projected/743a7318-33d0-4a59-93bd-7c6899554e5e-kube-api-access-fsdsv\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.808671 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.308656082 +0000 UTC m=+232.677282415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.809243 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-utilities\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.809253 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-catalog-content\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.842633 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsdsv\" (UniqueName: \"kubernetes.io/projected/743a7318-33d0-4a59-93bd-7c6899554e5e-kube-api-access-fsdsv\") pod \"certified-operators-zpnq8\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.846130 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.846848 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.854894 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.855059 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.858025 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.880698 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:31 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:31 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:31 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.880961 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.880721 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.910113 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:31 crc kubenswrapper[4926]: E0312 18:06:31.910464 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.410452116 +0000 UTC m=+232.779078449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.959387 4926 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.970243 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4zf5" event={"ID":"b2a609cd-c298-4356-9ddf-a7f125b52938","Type":"ContainerStarted","Data":"6da98836531143c0f0c9e5b0f54f53b5a8c52f32a98b1d55d2cd5e57c0ec7c9e"} Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.979535 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" event={"ID":"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a","Type":"ContainerStarted","Data":"5c2190b7dcff0a0981911cc6e9cc8e221d3f909aececbc7028daeebdaaddefca"} Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.979574 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" event={"ID":"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a","Type":"ContainerStarted","Data":"9b97c5c1e3245e013da7b8a53533ab149e2f2a6606fb877d381e7f0febac2c97"} Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.980071 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.986156 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" event={"ID":"8acca805-2bf8-4dcb-a036-c4732084210e","Type":"ContainerStarted","Data":"86d6b8eb4e0c5d613bb37d960f1cb609607647ef4a83a5fee4e0dc3cd1316553"} Mar 12 18:06:31 crc kubenswrapper[4926]: I0312 18:06:31.986198 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" event={"ID":"8acca805-2bf8-4dcb-a036-c4732084210e","Type":"ContainerStarted","Data":"945ca6bf80b6220cef8ef56354c13d5eb325de432c72b426448c9273ad83e0c1"} Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.003073 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-565fl"] Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.004670 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" podStartSLOduration=4.00464958 podStartE2EDuration="4.00464958s" podCreationTimestamp="2026-03-12 18:06:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:32.002521811 +0000 UTC m=+232.371148144" watchObservedRunningTime="2026-03-12 18:06:32.00464958 +0000 UTC m=+232.373275913" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.008482 4926 patch_prober.go:28] interesting pod/controller-manager-747cdc9dc9-hpk65 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" start-of-body= Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.008536 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" podUID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.012095 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.012528 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.012588 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.012754 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.512736855 +0000 UTC m=+232.881363188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.013338 4926 ???:1] "http: TLS handshake error from 192.168.126.11:38708: no serving certificate available for the kubelet" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.114065 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.114106 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.114196 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.118555 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.61853767 +0000 UTC m=+232.987164003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.119190 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.151831 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.199789 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.215766 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.216148 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.716133078 +0000 UTC m=+233.084759411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.313253 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t4sbh"] Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.317148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.317529 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.817508121 +0000 UTC m=+233.186134514 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.333258 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.420197 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.420394 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.920357544 +0000 UTC m=+233.288983877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.420644 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4edd042a-f910-49c2-9220-56a6e79b04dc-secret-volume\") pod \"4edd042a-f910-49c2-9220-56a6e79b04dc\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.420725 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4edd042a-f910-49c2-9220-56a6e79b04dc-config-volume\") pod \"4edd042a-f910-49c2-9220-56a6e79b04dc\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.420745 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq9lz\" (UniqueName: \"kubernetes.io/projected/4edd042a-f910-49c2-9220-56a6e79b04dc-kube-api-access-hq9lz\") pod \"4edd042a-f910-49c2-9220-56a6e79b04dc\" (UID: \"4edd042a-f910-49c2-9220-56a6e79b04dc\") " Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.420925 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.421302 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:32.92128586 +0000 UTC m=+233.289912193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.422270 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4edd042a-f910-49c2-9220-56a6e79b04dc-config-volume" (OuterVolumeSpecName: "config-volume") pod "4edd042a-f910-49c2-9220-56a6e79b04dc" (UID: "4edd042a-f910-49c2-9220-56a6e79b04dc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.432606 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpnq8"] Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.433344 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4edd042a-f910-49c2-9220-56a6e79b04dc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4edd042a-f910-49c2-9220-56a6e79b04dc" (UID: "4edd042a-f910-49c2-9220-56a6e79b04dc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.443296 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4edd042a-f910-49c2-9220-56a6e79b04dc-kube-api-access-hq9lz" (OuterVolumeSpecName: "kube-api-access-hq9lz") pod "4edd042a-f910-49c2-9220-56a6e79b04dc" (UID: "4edd042a-f910-49c2-9220-56a6e79b04dc"). InnerVolumeSpecName "kube-api-access-hq9lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.520603 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4230f869-9456-44a1-87b3-342fc8c18ed7" path="/var/lib/kubelet/pods/4230f869-9456-44a1-87b3-342fc8c18ed7/volumes" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.521298 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab2da4b6-fc75-4f97-b27c-3f687ddf9d98" path="/var/lib/kubelet/pods/ab2da4b6-fc75-4f97-b27c-3f687ddf9d98/volumes" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.522382 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.522833 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4edd042a-f910-49c2-9220-56a6e79b04dc-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.522854 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4edd042a-f910-49c2-9220-56a6e79b04dc-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.522866 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq9lz\" (UniqueName: \"kubernetes.io/projected/4edd042a-f910-49c2-9220-56a6e79b04dc-kube-api-access-hq9lz\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.522932 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:33.02291541 +0000 UTC m=+233.391541743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.523332 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 18:06:32 crc kubenswrapper[4926]: W0312 18:06:32.573428 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0156cb9a_b076_4cc2_9a14_68e06a5bbed4.slice/crio-c488202eea99f9ecf2c2a185d13fc76aef801aa5dd8c49cf087101de5890d5a4 WatchSource:0}: Error finding container c488202eea99f9ecf2c2a185d13fc76aef801aa5dd8c49cf087101de5890d5a4: Status 404 returned error can't find the container with id c488202eea99f9ecf2c2a185d13fc76aef801aa5dd8c49cf087101de5890d5a4 Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.624008 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.624398 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:33.124385815 +0000 UTC m=+233.493012148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.707586 4926 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-12T18:06:31.95942032Z","Handler":null,"Name":""} Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.725495 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.725784 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 18:06:33.225753638 +0000 UTC m=+233.594379971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.726721 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.727299 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 18:06:33.227276551 +0000 UTC m=+233.595902894 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g6fzt" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.736315 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs"] Mar 12 18:06:32 crc kubenswrapper[4926]: E0312 18:06:32.736680 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4edd042a-f910-49c2-9220-56a6e79b04dc" containerName="collect-profiles" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.736700 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4edd042a-f910-49c2-9220-56a6e79b04dc" containerName="collect-profiles" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.736830 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4edd042a-f910-49c2-9220-56a6e79b04dc" containerName="collect-profiles" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.737312 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.743822 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.744078 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.744713 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.744910 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.745052 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.745200 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.765452 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs"] Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.782345 4926 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.782383 4926 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.807562 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.808278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.821202 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.822017 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.824400 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.827284 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.827581 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb6dx\" (UniqueName: \"kubernetes.io/projected/7e957b87-36f8-433f-9547-70bf22354018-kube-api-access-xb6dx\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.827629 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-config\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.827657 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-client-ca\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.827703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e957b87-36f8-433f-9547-70bf22354018-serving-cert\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.871391 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.876927 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:32 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:32 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:32 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.876982 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.929845 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d00c7182-7f78-4452-a377-670193a7366b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.929906 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d00c7182-7f78-4452-a377-670193a7366b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.930051 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e957b87-36f8-433f-9547-70bf22354018-serving-cert\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.930285 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb6dx\" (UniqueName: \"kubernetes.io/projected/7e957b87-36f8-433f-9547-70bf22354018-kube-api-access-xb6dx\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.930369 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.930397 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-config\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.930475 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-client-ca\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.931885 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-client-ca\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.933046 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-config\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.937470 4926 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.937527 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.938588 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e957b87-36f8-433f-9547-70bf22354018-serving-cert\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.955261 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb6dx\" (UniqueName: \"kubernetes.io/projected/7e957b87-36f8-433f-9547-70bf22354018-kube-api-access-xb6dx\") pod \"route-controller-manager-758549c5dd-zclbs\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.995361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0156cb9a-b076-4cc2-9a14-68e06a5bbed4","Type":"ContainerStarted","Data":"c488202eea99f9ecf2c2a185d13fc76aef801aa5dd8c49cf087101de5890d5a4"} Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.998164 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" event={"ID":"4edd042a-f910-49c2-9220-56a6e79b04dc","Type":"ContainerDied","Data":"bf6d0c0b52a287d964339cdd9c78d27ee7b19ab7a4e80185b1b7ae31464e1f68"} Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.998189 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf6d0c0b52a287d964339cdd9c78d27ee7b19ab7a4e80185b1b7ae31464e1f68" Mar 12 18:06:32 crc kubenswrapper[4926]: I0312 18:06:32.998271 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555640-5965f" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.005309 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g6fzt\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.005381 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerID="fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8" exitCode=0 Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.005425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-565fl" event={"ID":"b5fe4032-6a1e-4c27-9471-fa53e044826e","Type":"ContainerDied","Data":"fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.005456 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-565fl" event={"ID":"b5fe4032-6a1e-4c27-9471-fa53e044826e","Type":"ContainerStarted","Data":"2ea4f4bc9037632e2b998da8312ddbb98d227eabc09259bd968b95eab6d3b562"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.016589 4926 generic.go:334] "Generic (PLEG): container finished" podID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerID="12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447" exitCode=0 Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.016999 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4zf5" event={"ID":"b2a609cd-c298-4356-9ddf-a7f125b52938","Type":"ContainerDied","Data":"12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.022217 4926 generic.go:334] "Generic (PLEG): container finished" podID="150781c8-5ae3-42a6-b351-2388dfe84167" containerID="2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139" exitCode=0 Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.022291 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t4sbh" event={"ID":"150781c8-5ae3-42a6-b351-2388dfe84167","Type":"ContainerDied","Data":"2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.022319 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t4sbh" event={"ID":"150781c8-5ae3-42a6-b351-2388dfe84167","Type":"ContainerStarted","Data":"f0734678dc103280ac23f9e4a0964bfd2dcfda20bb5ae8365cb587d30b175b62"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.031755 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d00c7182-7f78-4452-a377-670193a7366b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.031814 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d00c7182-7f78-4452-a377-670193a7366b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.031929 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d00c7182-7f78-4452-a377-670193a7366b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.037882 4926 generic.go:334] "Generic (PLEG): container finished" podID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerID="80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9" exitCode=0 Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.038714 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpnq8" event={"ID":"743a7318-33d0-4a59-93bd-7c6899554e5e","Type":"ContainerDied","Data":"80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.038766 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpnq8" event={"ID":"743a7318-33d0-4a59-93bd-7c6899554e5e","Type":"ContainerStarted","Data":"dc92306c9f92008ccb5e4b67f4e37a7020d96c316ccb9bc2ecf40d9ac23d9cf2"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.061590 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d00c7182-7f78-4452-a377-670193a7366b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.062562 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" event={"ID":"8acca805-2bf8-4dcb-a036-c4732084210e","Type":"ContainerStarted","Data":"72c3731c75f68b012ce99f7bd7ad25d84a188d56b928b2ed7e8919fba03a328b"} Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.065640 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.070142 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.125106 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-c9xl5" podStartSLOduration=11.125089108 podStartE2EDuration="11.125089108s" podCreationTimestamp="2026-03-12 18:06:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:33.120807018 +0000 UTC m=+233.489433351" watchObservedRunningTime="2026-03-12 18:06:33.125089108 +0000 UTC m=+233.493715441" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.154236 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wz6qt"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.155811 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.155941 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.161179 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.194942 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz6qt"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.200762 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.247245 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-utilities\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.247871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-catalog-content\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.248045 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwlr2\" (UniqueName: \"kubernetes.io/projected/702daa2d-851e-4c3d-be86-4f337b4462f7-kube-api-access-nwlr2\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.364552 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-utilities\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.364645 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-catalog-content\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.364709 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwlr2\" (UniqueName: \"kubernetes.io/projected/702daa2d-851e-4c3d-be86-4f337b4462f7-kube-api-access-nwlr2\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.365718 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-catalog-content\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.365919 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-utilities\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.391228 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwlr2\" (UniqueName: \"kubernetes.io/projected/702daa2d-851e-4c3d-be86-4f337b4462f7-kube-api-access-nwlr2\") pod \"redhat-marketplace-wz6qt\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.553856 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.564080 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v5qw8"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.565365 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.587561 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5qw8"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.676271 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-catalog-content\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.676327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-utilities\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.676381 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkgrh\" (UniqueName: \"kubernetes.io/projected/8630848f-c268-4f4a-9fd0-8f33765c20b4-kube-api-access-pkgrh\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.688058 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.743763 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.762209 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g6fzt"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.777900 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-catalog-content\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.777951 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-utilities\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.777998 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkgrh\" (UniqueName: \"kubernetes.io/projected/8630848f-c268-4f4a-9fd0-8f33765c20b4-kube-api-access-pkgrh\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.778649 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-catalog-content\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.778851 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-utilities\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.814967 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkgrh\" (UniqueName: \"kubernetes.io/projected/8630848f-c268-4f4a-9fd0-8f33765c20b4-kube-api-access-pkgrh\") pod \"redhat-marketplace-v5qw8\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.863327 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:33 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:33 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:33 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.863717 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.903986 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.949689 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz6qt"] Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.999284 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:33 crc kubenswrapper[4926]: I0312 18:06:33.999343 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.005626 4926 patch_prober.go:28] interesting pod/console-f9d7485db-vb9qx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.005764 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vb9qx" podUID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 12 18:06:34 crc kubenswrapper[4926]: W0312 18:06:34.009708 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod702daa2d_851e_4c3d_be86_4f337b4462f7.slice/crio-1ca7001bbffd391adca6fddca7abaa8f8e5f4a6dec4a426455ef61c2d31fac2d WatchSource:0}: Error finding container 1ca7001bbffd391adca6fddca7abaa8f8e5f4a6dec4a426455ef61c2d31fac2d: Status 404 returned error can't find the container with id 1ca7001bbffd391adca6fddca7abaa8f8e5f4a6dec4a426455ef61c2d31fac2d Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.079600 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" event={"ID":"7e957b87-36f8-433f-9547-70bf22354018","Type":"ContainerStarted","Data":"9f487eeed2150c2eeef5da93b9c67da82e7c33df3d94484e998aa4c72eec6d5a"} Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.081638 4926 generic.go:334] "Generic (PLEG): container finished" podID="0156cb9a-b076-4cc2-9a14-68e06a5bbed4" containerID="a9d8fd89ceed81099f6030ffc1f3c040d3c1980ae97507aa4cd5759f3a0ff3d1" exitCode=0 Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.081692 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0156cb9a-b076-4cc2-9a14-68e06a5bbed4","Type":"ContainerDied","Data":"a9d8fd89ceed81099f6030ffc1f3c040d3c1980ae97507aa4cd5759f3a0ff3d1"} Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.083936 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d00c7182-7f78-4452-a377-670193a7366b","Type":"ContainerStarted","Data":"bca04b15f8cd7489625ca0ac157b17db2f261392d7437acf647da36fff99051b"} Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.109541 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" event={"ID":"97b0faa2-bcb2-417e-9065-3156860a8644","Type":"ContainerStarted","Data":"5fd53c65057b16b866539f523a9ef035c2f73d7900692a8369fc22f3d2f8f197"} Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.111009 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz6qt" event={"ID":"702daa2d-851e-4c3d-be86-4f337b4462f7","Type":"ContainerStarted","Data":"1ca7001bbffd391adca6fddca7abaa8f8e5f4a6dec4a426455ef61c2d31fac2d"} Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.128768 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.128803 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.144630 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.151861 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zx9h7"] Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.152852 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.156605 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.186054 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zx9h7"] Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.223672 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.233992 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9bzkt" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.294892 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-catalog-content\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.295086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57ncv\" (UniqueName: \"kubernetes.io/projected/637236a6-6287-401d-a2cd-78713aa03176-kube-api-access-57ncv\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.295662 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-utilities\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.397288 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.397379 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57ncv\" (UniqueName: \"kubernetes.io/projected/637236a6-6287-401d-a2cd-78713aa03176-kube-api-access-57ncv\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.397413 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.397476 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-utilities\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.397534 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-catalog-content\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.398002 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-catalog-content\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.399450 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-utilities\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.409405 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.421189 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.448482 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5qw8"] Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.451281 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57ncv\" (UniqueName: \"kubernetes.io/projected/637236a6-6287-401d-a2cd-78713aa03176-kube-api-access-57ncv\") pod \"redhat-operators-zx9h7\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.503851 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.504120 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.504182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.510290 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.513776 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.522879 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.538992 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.548065 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.562950 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hmn6x"] Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.563991 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.587515 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmn6x"] Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.720411 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-catalog-content\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.720773 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4msz\" (UniqueName: \"kubernetes.io/projected/1f425571-9ce5-4fdc-9631-7683efa292aa-kube-api-access-v4msz\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.720802 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-utilities\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.807166 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.822377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-catalog-content\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.822427 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4msz\" (UniqueName: \"kubernetes.io/projected/1f425571-9ce5-4fdc-9631-7683efa292aa-kube-api-access-v4msz\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.822467 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-utilities\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.822890 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-utilities\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.822938 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-catalog-content\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.862249 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4msz\" (UniqueName: \"kubernetes.io/projected/1f425571-9ce5-4fdc-9631-7683efa292aa-kube-api-access-v4msz\") pod \"redhat-operators-hmn6x\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.865551 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.880643 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:34 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:34 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:34 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.880695 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:34 crc kubenswrapper[4926]: I0312 18:06:34.896559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.133895 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" event={"ID":"7e957b87-36f8-433f-9547-70bf22354018","Type":"ContainerStarted","Data":"404fc2dbd169e6a506a47a16f7ea299950a21ae0b3144c3fe7d945d26a749973"} Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.134392 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.139710 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.148077 4926 generic.go:334] "Generic (PLEG): container finished" podID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerID="ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3" exitCode=0 Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.148145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5qw8" event={"ID":"8630848f-c268-4f4a-9fd0-8f33765c20b4","Type":"ContainerDied","Data":"ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3"} Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.148165 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5qw8" event={"ID":"8630848f-c268-4f4a-9fd0-8f33765c20b4","Type":"ContainerStarted","Data":"e7bb37ba860cfb3ecfc47e3421fdc163d8b676f30a5d56c59a6b61ad57e2cd0c"} Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.152533 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" podStartSLOduration=7.152521569 podStartE2EDuration="7.152521569s" podCreationTimestamp="2026-03-12 18:06:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:35.147420997 +0000 UTC m=+235.516047330" watchObservedRunningTime="2026-03-12 18:06:35.152521569 +0000 UTC m=+235.521147902" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.159693 4926 generic.go:334] "Generic (PLEG): container finished" podID="d00c7182-7f78-4452-a377-670193a7366b" containerID="01d6a31eb39da97367417192915f91b07b207f5bf4fbb0c3e97c4026c48a7c65" exitCode=0 Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.159751 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d00c7182-7f78-4452-a377-670193a7366b","Type":"ContainerDied","Data":"01d6a31eb39da97367417192915f91b07b207f5bf4fbb0c3e97c4026c48a7c65"} Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.174921 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" event={"ID":"97b0faa2-bcb2-417e-9065-3156860a8644","Type":"ContainerStarted","Data":"f7c989898f3b81f561b1ebddb62bcb7b02ee6fac8bd401ab3a5ff75b154fda11"} Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.175227 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.188263 4926 generic.go:334] "Generic (PLEG): container finished" podID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerID="7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa" exitCode=0 Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.189896 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz6qt" event={"ID":"702daa2d-851e-4c3d-be86-4f337b4462f7","Type":"ContainerDied","Data":"7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa"} Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.193962 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-z95pp" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.230407 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" podStartSLOduration=164.230390148 podStartE2EDuration="2m44.230390148s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:06:35.229234735 +0000 UTC m=+235.597861068" watchObservedRunningTime="2026-03-12 18:06:35.230390148 +0000 UTC m=+235.599016481" Mar 12 18:06:35 crc kubenswrapper[4926]: W0312 18:06:35.435370 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f741f51ff41e6e41540c659345accd524bff2e736d76e15c0688827a3209d685 WatchSource:0}: Error finding container f741f51ff41e6e41540c659345accd524bff2e736d76e15c0688827a3209d685: Status 404 returned error can't find the container with id f741f51ff41e6e41540c659345accd524bff2e736d76e15c0688827a3209d685 Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.503364 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmn6x"] Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.605762 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zx9h7"] Mar 12 18:06:35 crc kubenswrapper[4926]: W0312 18:06:35.616383 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod637236a6_6287_401d_a2cd_78713aa03176.slice/crio-b41dbdfd44a141e983428f384fe92d330459deb4bd26d614487e897aef5edb00 WatchSource:0}: Error finding container b41dbdfd44a141e983428f384fe92d330459deb4bd26d614487e897aef5edb00: Status 404 returned error can't find the container with id b41dbdfd44a141e983428f384fe92d330459deb4bd26d614487e897aef5edb00 Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.671923 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-st67x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.671960 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-st67x" podUID="06394469-defd-4710-90a2-b6c395c00d4f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.672064 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-st67x container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.672111 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-st67x" podUID="06394469-defd-4710-90a2-b6c395c00d4f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.706261 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.854726 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kubelet-dir\") pod \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.855111 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0156cb9a-b076-4cc2-9a14-68e06a5bbed4" (UID: "0156cb9a-b076-4cc2-9a14-68e06a5bbed4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.855563 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kube-api-access\") pod \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\" (UID: \"0156cb9a-b076-4cc2-9a14-68e06a5bbed4\") " Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.856125 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.865148 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:35 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:35 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:35 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.865225 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.891848 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0156cb9a-b076-4cc2-9a14-68e06a5bbed4" (UID: "0156cb9a-b076-4cc2-9a14-68e06a5bbed4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:06:35 crc kubenswrapper[4926]: I0312 18:06:35.957456 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0156cb9a-b076-4cc2-9a14-68e06a5bbed4-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.224896 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"74230acbcb2ff2532edb6856289695dc59296f47c7f8b82ad1bac248d06f52e3"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.224939 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"1f4b5d0934532a9a6ba9d33f6a6ad066d61dd50d39fc0e2594fed50d2467c7d8"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.225132 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.230273 4926 generic.go:334] "Generic (PLEG): container finished" podID="637236a6-6287-401d-a2cd-78713aa03176" containerID="76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1" exitCode=0 Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.230358 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx9h7" event={"ID":"637236a6-6287-401d-a2cd-78713aa03176","Type":"ContainerDied","Data":"76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.230424 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx9h7" event={"ID":"637236a6-6287-401d-a2cd-78713aa03176","Type":"ContainerStarted","Data":"b41dbdfd44a141e983428f384fe92d330459deb4bd26d614487e897aef5edb00"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.242259 4926 generic.go:334] "Generic (PLEG): container finished" podID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerID="3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a" exitCode=0 Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.242644 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmn6x" event={"ID":"1f425571-9ce5-4fdc-9631-7683efa292aa","Type":"ContainerDied","Data":"3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.242731 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmn6x" event={"ID":"1f425571-9ce5-4fdc-9631-7683efa292aa","Type":"ContainerStarted","Data":"d295389691829deab9da8e29d0c26dfd809e13bacff8864270d0d1a8c5b14dcd"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.253425 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.246958 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"bc5d581fe3ff1e668de4215b5f77628054188baca4bf123fd67c0191c436b563"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.254049 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fa1688dc92d10fddca52e3788554a0fe6becc03f8ec0c58b1997076956d0a899"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.254070 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1447bf24555e677d095d00ea7874b121bd3cb802eddb13ee5c1823a91ae9712c"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.254084 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f741f51ff41e6e41540c659345accd524bff2e736d76e15c0688827a3209d685"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.254097 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"0156cb9a-b076-4cc2-9a14-68e06a5bbed4","Type":"ContainerDied","Data":"c488202eea99f9ecf2c2a185d13fc76aef801aa5dd8c49cf087101de5890d5a4"} Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.254116 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c488202eea99f9ecf2c2a185d13fc76aef801aa5dd8c49cf087101de5890d5a4" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.700596 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.773121 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d00c7182-7f78-4452-a377-670193a7366b-kubelet-dir\") pod \"d00c7182-7f78-4452-a377-670193a7366b\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.773225 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d00c7182-7f78-4452-a377-670193a7366b-kube-api-access\") pod \"d00c7182-7f78-4452-a377-670193a7366b\" (UID: \"d00c7182-7f78-4452-a377-670193a7366b\") " Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.773511 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d00c7182-7f78-4452-a377-670193a7366b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d00c7182-7f78-4452-a377-670193a7366b" (UID: "d00c7182-7f78-4452-a377-670193a7366b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.780480 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d00c7182-7f78-4452-a377-670193a7366b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d00c7182-7f78-4452-a377-670193a7366b" (UID: "d00c7182-7f78-4452-a377-670193a7366b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.875208 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d00c7182-7f78-4452-a377-670193a7366b-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.875246 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d00c7182-7f78-4452-a377-670193a7366b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.878346 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:36 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:36 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:36 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:36 crc kubenswrapper[4926]: I0312 18:06:36.878402 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:37 crc kubenswrapper[4926]: I0312 18:06:37.188145 4926 ???:1] "http: TLS handshake error from 192.168.126.11:50054: no serving certificate available for the kubelet" Mar 12 18:06:37 crc kubenswrapper[4926]: I0312 18:06:37.316661 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d00c7182-7f78-4452-a377-670193a7366b","Type":"ContainerDied","Data":"bca04b15f8cd7489625ca0ac157b17db2f261392d7437acf647da36fff99051b"} Mar 12 18:06:37 crc kubenswrapper[4926]: I0312 18:06:37.316742 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bca04b15f8cd7489625ca0ac157b17db2f261392d7437acf647da36fff99051b" Mar 12 18:06:37 crc kubenswrapper[4926]: I0312 18:06:37.316787 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 18:06:37 crc kubenswrapper[4926]: I0312 18:06:37.865778 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:37 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:37 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:37 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:37 crc kubenswrapper[4926]: I0312 18:06:37.865840 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:38 crc kubenswrapper[4926]: I0312 18:06:38.865377 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:38 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:38 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:38 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:38 crc kubenswrapper[4926]: I0312 18:06:38.865467 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:39 crc kubenswrapper[4926]: I0312 18:06:39.864810 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:39 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:39 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:39 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:39 crc kubenswrapper[4926]: I0312 18:06:39.865154 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:40 crc kubenswrapper[4926]: I0312 18:06:40.220909 4926 ???:1] "http: TLS handshake error from 192.168.126.11:50064: no serving certificate available for the kubelet" Mar 12 18:06:40 crc kubenswrapper[4926]: I0312 18:06:40.594709 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-rxml7" Mar 12 18:06:40 crc kubenswrapper[4926]: I0312 18:06:40.871320 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:40 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:40 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:40 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:40 crc kubenswrapper[4926]: I0312 18:06:40.871383 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:41 crc kubenswrapper[4926]: I0312 18:06:41.862295 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:41 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:41 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:41 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:41 crc kubenswrapper[4926]: I0312 18:06:41.862377 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:42 crc kubenswrapper[4926]: I0312 18:06:42.862855 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:42 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:42 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:42 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:42 crc kubenswrapper[4926]: I0312 18:06:42.862915 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:43 crc kubenswrapper[4926]: I0312 18:06:43.862781 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:43 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:43 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:43 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:43 crc kubenswrapper[4926]: I0312 18:06:43.863044 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:43 crc kubenswrapper[4926]: I0312 18:06:43.999648 4926 patch_prober.go:28] interesting pod/console-f9d7485db-vb9qx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 12 18:06:43 crc kubenswrapper[4926]: I0312 18:06:43.999842 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vb9qx" podUID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.25:8443/health\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 12 18:06:44 crc kubenswrapper[4926]: I0312 18:06:44.801088 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:44 crc kubenswrapper[4926]: I0312 18:06:44.803065 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 18:06:44 crc kubenswrapper[4926]: I0312 18:06:44.820277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/211eeae6-9b41-484b-bd13-99c1c28cdf96-metrics-certs\") pod \"network-metrics-daemon-n7pd7\" (UID: \"211eeae6-9b41-484b-bd13-99c1c28cdf96\") " pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:44 crc kubenswrapper[4926]: I0312 18:06:44.862497 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:44 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Mar 12 18:06:44 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:44 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:44 crc kubenswrapper[4926]: I0312 18:06:44.862547 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.015997 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.025158 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-n7pd7" Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.671719 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-st67x container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.671780 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-st67x container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.671831 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-st67x" podUID="06394469-defd-4710-90a2-b6c395c00d4f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.671779 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-st67x" podUID="06394469-defd-4710-90a2-b6c395c00d4f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.863018 4926 patch_prober.go:28] interesting pod/router-default-5444994796-wndvq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 18:06:45 crc kubenswrapper[4926]: [+]has-synced ok Mar 12 18:06:45 crc kubenswrapper[4926]: [+]process-running ok Mar 12 18:06:45 crc kubenswrapper[4926]: healthz check failed Mar 12 18:06:45 crc kubenswrapper[4926]: I0312 18:06:45.863101 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-wndvq" podUID="e379fe1d-7780-4f17-8df8-f74f3dddbc23" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 18:06:46 crc kubenswrapper[4926]: I0312 18:06:46.863456 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:46 crc kubenswrapper[4926]: I0312 18:06:46.866604 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-wndvq" Mar 12 18:06:47 crc kubenswrapper[4926]: I0312 18:06:47.135597 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-747cdc9dc9-hpk65"] Mar 12 18:06:47 crc kubenswrapper[4926]: I0312 18:06:47.136230 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" podUID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" containerName="controller-manager" containerID="cri-o://5c2190b7dcff0a0981911cc6e9cc8e221d3f909aececbc7028daeebdaaddefca" gracePeriod=30 Mar 12 18:06:47 crc kubenswrapper[4926]: I0312 18:06:47.143361 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs"] Mar 12 18:06:47 crc kubenswrapper[4926]: I0312 18:06:47.143589 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" podUID="7e957b87-36f8-433f-9547-70bf22354018" containerName="route-controller-manager" containerID="cri-o://404fc2dbd169e6a506a47a16f7ea299950a21ae0b3144c3fe7d945d26a749973" gracePeriod=30 Mar 12 18:06:47 crc kubenswrapper[4926]: I0312 18:06:47.462471 4926 ???:1] "http: TLS handshake error from 192.168.126.11:44194: no serving certificate available for the kubelet" Mar 12 18:06:48 crc kubenswrapper[4926]: I0312 18:06:48.465197 4926 generic.go:334] "Generic (PLEG): container finished" podID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" containerID="5c2190b7dcff0a0981911cc6e9cc8e221d3f909aececbc7028daeebdaaddefca" exitCode=0 Mar 12 18:06:48 crc kubenswrapper[4926]: I0312 18:06:48.465295 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" event={"ID":"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a","Type":"ContainerDied","Data":"5c2190b7dcff0a0981911cc6e9cc8e221d3f909aececbc7028daeebdaaddefca"} Mar 12 18:06:48 crc kubenswrapper[4926]: I0312 18:06:48.467913 4926 generic.go:334] "Generic (PLEG): container finished" podID="7e957b87-36f8-433f-9547-70bf22354018" containerID="404fc2dbd169e6a506a47a16f7ea299950a21ae0b3144c3fe7d945d26a749973" exitCode=0 Mar 12 18:06:48 crc kubenswrapper[4926]: I0312 18:06:48.467987 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" event={"ID":"7e957b87-36f8-433f-9547-70bf22354018","Type":"ContainerDied","Data":"404fc2dbd169e6a506a47a16f7ea299950a21ae0b3144c3fe7d945d26a749973"} Mar 12 18:06:51 crc kubenswrapper[4926]: I0312 18:06:51.033479 4926 patch_prober.go:28] interesting pod/controller-manager-747cdc9dc9-hpk65 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" start-of-body= Mar 12 18:06:51 crc kubenswrapper[4926]: I0312 18:06:51.033599 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" podUID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.45:8443/healthz\": dial tcp 10.217.0.45:8443: connect: connection refused" Mar 12 18:06:53 crc kubenswrapper[4926]: I0312 18:06:53.067125 4926 patch_prober.go:28] interesting pod/route-controller-manager-758549c5dd-zclbs container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" start-of-body= Mar 12 18:06:53 crc kubenswrapper[4926]: I0312 18:06:53.067429 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" podUID="7e957b87-36f8-433f-9547-70bf22354018" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": dial tcp 10.217.0.51:8443: connect: connection refused" Mar 12 18:06:53 crc kubenswrapper[4926]: I0312 18:06:53.208779 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:06:54 crc kubenswrapper[4926]: I0312 18:06:54.003893 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:54 crc kubenswrapper[4926]: I0312 18:06:54.007916 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:06:55 crc kubenswrapper[4926]: I0312 18:06:55.684963 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-st67x" Mar 12 18:06:56 crc kubenswrapper[4926]: I0312 18:06:56.818132 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:06:56 crc kubenswrapper[4926]: I0312 18:06:56.818220 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:06:57 crc kubenswrapper[4926]: E0312 18:06:57.530536 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 12 18:06:57 crc kubenswrapper[4926]: E0312 18:06:57.530975 4926 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 12 18:06:57 crc kubenswrapper[4926]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 12 18:06:57 crc kubenswrapper[4926]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tl7pp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29555646-wqpkb_openshift-infra(d68160cf-4e6c-4294-bfdc-4acb74637ecb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 12 18:06:57 crc kubenswrapper[4926]: > logger="UnhandledError" Mar 12 18:06:57 crc kubenswrapper[4926]: E0312 18:06:57.532121 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" podUID="d68160cf-4e6c-4294-bfdc-4acb74637ecb" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.934379 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.945691 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.964686 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx"] Mar 12 18:06:57 crc kubenswrapper[4926]: E0312 18:06:57.964945 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e957b87-36f8-433f-9547-70bf22354018" containerName="route-controller-manager" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.964960 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e957b87-36f8-433f-9547-70bf22354018" containerName="route-controller-manager" Mar 12 18:06:57 crc kubenswrapper[4926]: E0312 18:06:57.964981 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d00c7182-7f78-4452-a377-670193a7366b" containerName="pruner" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.964990 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d00c7182-7f78-4452-a377-670193a7366b" containerName="pruner" Mar 12 18:06:57 crc kubenswrapper[4926]: E0312 18:06:57.965005 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0156cb9a-b076-4cc2-9a14-68e06a5bbed4" containerName="pruner" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.965013 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0156cb9a-b076-4cc2-9a14-68e06a5bbed4" containerName="pruner" Mar 12 18:06:57 crc kubenswrapper[4926]: E0312 18:06:57.965029 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" containerName="controller-manager" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.965037 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" containerName="controller-manager" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.965151 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0156cb9a-b076-4cc2-9a14-68e06a5bbed4" containerName="pruner" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.965171 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d00c7182-7f78-4452-a377-670193a7366b" containerName="pruner" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.965186 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e957b87-36f8-433f-9547-70bf22354018" containerName="route-controller-manager" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.965195 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" containerName="controller-manager" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.965615 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:57 crc kubenswrapper[4926]: I0312 18:06:57.980030 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx"] Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.013223 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb6dx\" (UniqueName: \"kubernetes.io/projected/7e957b87-36f8-433f-9547-70bf22354018-kube-api-access-xb6dx\") pod \"7e957b87-36f8-433f-9547-70bf22354018\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.014946 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-client-ca\") pod \"7e957b87-36f8-433f-9547-70bf22354018\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.015116 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hbtx\" (UniqueName: \"kubernetes.io/projected/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-kube-api-access-5hbtx\") pod \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.015161 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-config\") pod \"7e957b87-36f8-433f-9547-70bf22354018\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.015957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-client-ca" (OuterVolumeSpecName: "client-ca") pod "7e957b87-36f8-433f-9547-70bf22354018" (UID: "7e957b87-36f8-433f-9547-70bf22354018"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.020424 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-config" (OuterVolumeSpecName: "config") pod "7e957b87-36f8-433f-9547-70bf22354018" (UID: "7e957b87-36f8-433f-9547-70bf22354018"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.021156 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-kube-api-access-5hbtx" (OuterVolumeSpecName: "kube-api-access-5hbtx") pod "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" (UID: "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a"). InnerVolumeSpecName "kube-api-access-5hbtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.023504 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e957b87-36f8-433f-9547-70bf22354018-kube-api-access-xb6dx" (OuterVolumeSpecName: "kube-api-access-xb6dx") pod "7e957b87-36f8-433f-9547-70bf22354018" (UID: "7e957b87-36f8-433f-9547-70bf22354018"). InnerVolumeSpecName "kube-api-access-xb6dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.023698 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-config\") pod \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.023784 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-serving-cert\") pod \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.024998 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-config" (OuterVolumeSpecName: "config") pod "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" (UID: "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.025086 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-client-ca\") pod \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.025113 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-proxy-ca-bundles\") pod \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\" (UID: \"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.025715 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e957b87-36f8-433f-9547-70bf22354018-serving-cert\") pod \"7e957b87-36f8-433f-9547-70bf22354018\" (UID: \"7e957b87-36f8-433f-9547-70bf22354018\") " Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.025765 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" (UID: "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026206 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxl78\" (UniqueName: \"kubernetes.io/projected/91f58aee-72fa-449f-8ec9-c80f89e5bc39-kube-api-access-jxl78\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026430 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-client-ca" (OuterVolumeSpecName: "client-ca") pod "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" (UID: "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026567 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-config\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026641 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91f58aee-72fa-449f-8ec9-c80f89e5bc39-serving-cert\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026692 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-client-ca\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026808 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb6dx\" (UniqueName: \"kubernetes.io/projected/7e957b87-36f8-433f-9547-70bf22354018-kube-api-access-xb6dx\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026825 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026838 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hbtx\" (UniqueName: \"kubernetes.io/projected/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-kube-api-access-5hbtx\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026849 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e957b87-36f8-433f-9547-70bf22354018-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026860 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026871 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.026881 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.028254 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e957b87-36f8-433f-9547-70bf22354018-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7e957b87-36f8-433f-9547-70bf22354018" (UID: "7e957b87-36f8-433f-9547-70bf22354018"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.028380 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" (UID: "8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.128276 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxl78\" (UniqueName: \"kubernetes.io/projected/91f58aee-72fa-449f-8ec9-c80f89e5bc39-kube-api-access-jxl78\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.128387 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-config\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.128430 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91f58aee-72fa-449f-8ec9-c80f89e5bc39-serving-cert\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.128473 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-client-ca\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.128522 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e957b87-36f8-433f-9547-70bf22354018-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.128536 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.129453 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-client-ca\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.129716 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-config\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.133095 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91f58aee-72fa-449f-8ec9-c80f89e5bc39-serving-cert\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.144609 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxl78\" (UniqueName: \"kubernetes.io/projected/91f58aee-72fa-449f-8ec9-c80f89e5bc39-kube-api-access-jxl78\") pod \"route-controller-manager-b6dbb7676-l48qx\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.298967 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.530313 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" event={"ID":"8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a","Type":"ContainerDied","Data":"9b97c5c1e3245e013da7b8a53533ab149e2f2a6606fb877d381e7f0febac2c97"} Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.530336 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747cdc9dc9-hpk65" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.530368 4926 scope.go:117] "RemoveContainer" containerID="5c2190b7dcff0a0981911cc6e9cc8e221d3f909aececbc7028daeebdaaddefca" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.532877 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.533157 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs" event={"ID":"7e957b87-36f8-433f-9547-70bf22354018","Type":"ContainerDied","Data":"9f487eeed2150c2eeef5da93b9c67da82e7c33df3d94484e998aa4c72eec6d5a"} Mar 12 18:06:58 crc kubenswrapper[4926]: E0312 18:06:58.534068 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" podUID="d68160cf-4e6c-4294-bfdc-4acb74637ecb" Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.551131 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-747cdc9dc9-hpk65"] Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.553570 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-747cdc9dc9-hpk65"] Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.570178 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs"] Mar 12 18:06:58 crc kubenswrapper[4926]: I0312 18:06:58.572575 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-758549c5dd-zclbs"] Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.496204 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e957b87-36f8-433f-9547-70bf22354018" path="/var/lib/kubelet/pods/7e957b87-36f8-433f-9547-70bf22354018/volumes" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.497057 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a" path="/var/lib/kubelet/pods/8fe6b3ba-075b-4f7f-bc18-f6e4c1e4713a/volumes" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.767601 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c9659bb69-k2bhk"] Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.768761 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.773342 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.774160 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.775317 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.775960 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.776025 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.776034 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.777167 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c9659bb69-k2bhk"] Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.787064 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.870184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-client-ca\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.870227 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-proxy-ca-bundles\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.870260 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-config\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.870280 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f96e9f-9853-4f3b-8c84-3ba79491d133-serving-cert\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.870300 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49sdm\" (UniqueName: \"kubernetes.io/projected/23f96e9f-9853-4f3b-8c84-3ba79491d133-kube-api-access-49sdm\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.971568 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-client-ca\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.971845 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-proxy-ca-bundles\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.971869 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-config\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.971890 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f96e9f-9853-4f3b-8c84-3ba79491d133-serving-cert\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.971911 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49sdm\" (UniqueName: \"kubernetes.io/projected/23f96e9f-9853-4f3b-8c84-3ba79491d133-kube-api-access-49sdm\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.972527 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-client-ca\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.973325 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-proxy-ca-bundles\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.974594 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-config\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.976786 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f96e9f-9853-4f3b-8c84-3ba79491d133-serving-cert\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:00 crc kubenswrapper[4926]: I0312 18:07:00.991750 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49sdm\" (UniqueName: \"kubernetes.io/projected/23f96e9f-9853-4f3b-8c84-3ba79491d133-kube-api-access-49sdm\") pod \"controller-manager-7c9659bb69-k2bhk\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:01 crc kubenswrapper[4926]: I0312 18:07:01.106831 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:02 crc kubenswrapper[4926]: E0312 18:07:02.145034 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 12 18:07:02 crc kubenswrapper[4926]: E0312 18:07:02.145494 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tnjm4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-t4sbh_openshift-marketplace(150781c8-5ae3-42a6-b351-2388dfe84167): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:02 crc kubenswrapper[4926]: E0312 18:07:02.146708 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-t4sbh" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" Mar 12 18:07:02 crc kubenswrapper[4926]: E0312 18:07:02.147851 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 12 18:07:02 crc kubenswrapper[4926]: E0312 18:07:02.147989 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wf7tf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-s4zf5_openshift-marketplace(b2a609cd-c298-4356-9ddf-a7f125b52938): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:02 crc kubenswrapper[4926]: E0312 18:07:02.149555 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-s4zf5" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" Mar 12 18:07:03 crc kubenswrapper[4926]: E0312 18:07:03.683219 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-s4zf5" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" Mar 12 18:07:03 crc kubenswrapper[4926]: E0312 18:07:03.683225 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-t4sbh" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" Mar 12 18:07:03 crc kubenswrapper[4926]: E0312 18:07:03.738297 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 12 18:07:03 crc kubenswrapper[4926]: E0312 18:07:03.738536 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pkgrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-v5qw8_openshift-marketplace(8630848f-c268-4f4a-9fd0-8f33765c20b4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:03 crc kubenswrapper[4926]: E0312 18:07:03.739746 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-v5qw8" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" Mar 12 18:07:04 crc kubenswrapper[4926]: I0312 18:07:04.792112 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7rmsn"] Mar 12 18:07:05 crc kubenswrapper[4926]: I0312 18:07:05.885379 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jfb58" Mar 12 18:07:06 crc kubenswrapper[4926]: I0312 18:07:06.828386 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 18:07:06 crc kubenswrapper[4926]: I0312 18:07:06.829314 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:06 crc kubenswrapper[4926]: I0312 18:07:06.831301 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 12 18:07:06 crc kubenswrapper[4926]: I0312 18:07:06.831810 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 12 18:07:06 crc kubenswrapper[4926]: I0312 18:07:06.834525 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 18:07:06 crc kubenswrapper[4926]: I0312 18:07:06.945113 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51ec2202-9e22-4586-af8c-020d00160c26-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:06 crc kubenswrapper[4926]: I0312 18:07:06.945167 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51ec2202-9e22-4586-af8c-020d00160c26-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.046539 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51ec2202-9e22-4586-af8c-020d00160c26-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.046613 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51ec2202-9e22-4586-af8c-020d00160c26-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.046722 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51ec2202-9e22-4586-af8c-020d00160c26-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.068507 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51ec2202-9e22-4586-af8c-020d00160c26-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.133745 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9659bb69-k2bhk"] Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.150266 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.229618 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx"] Mar 12 18:07:07 crc kubenswrapper[4926]: E0312 18:07:07.784665 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-v5qw8" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" Mar 12 18:07:07 crc kubenswrapper[4926]: E0312 18:07:07.881650 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 12 18:07:07 crc kubenswrapper[4926]: E0312 18:07:07.881806 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v4msz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-hmn6x_openshift-marketplace(1f425571-9ce5-4fdc-9631-7683efa292aa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:07 crc kubenswrapper[4926]: E0312 18:07:07.883022 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-hmn6x" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" Mar 12 18:07:07 crc kubenswrapper[4926]: I0312 18:07:07.978606 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-n7pd7"] Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.577255 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-hmn6x" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" Mar 12 18:07:09 crc kubenswrapper[4926]: I0312 18:07:09.587286 4926 scope.go:117] "RemoveContainer" containerID="404fc2dbd169e6a506a47a16f7ea299950a21ae0b3144c3fe7d945d26a749973" Mar 12 18:07:09 crc kubenswrapper[4926]: W0312 18:07:09.608994 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod211eeae6_9b41_484b_bd13_99c1c28cdf96.slice/crio-fa523477acd29a14cfd122e3283db6ab6e928eb05746f29ea2b64f89366a97aa WatchSource:0}: Error finding container fa523477acd29a14cfd122e3283db6ab6e928eb05746f29ea2b64f89366a97aa: Status 404 returned error can't find the container with id fa523477acd29a14cfd122e3283db6ab6e928eb05746f29ea2b64f89366a97aa Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.668480 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.668610 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nwlr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wz6qt_openshift-marketplace(702daa2d-851e-4c3d-be86-4f337b4462f7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.669925 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-wz6qt" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.670597 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.670670 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2wxfn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-565fl_openshift-marketplace(b5fe4032-6a1e-4c27-9471-fa53e044826e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.671729 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-565fl" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.699590 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.699925 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-57ncv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zx9h7_openshift-marketplace(637236a6-6287-401d-a2cd-78713aa03176): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.701094 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zx9h7" podUID="637236a6-6287-401d-a2cd-78713aa03176" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.718869 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.719031 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fsdsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zpnq8_openshift-marketplace(743a7318-33d0-4a59-93bd-7c6899554e5e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 18:07:09 crc kubenswrapper[4926]: E0312 18:07:09.720375 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zpnq8" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" Mar 12 18:07:09 crc kubenswrapper[4926]: I0312 18:07:09.858572 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx"] Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.132680 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.138769 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9659bb69-k2bhk"] Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.594623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" event={"ID":"211eeae6-9b41-484b-bd13-99c1c28cdf96","Type":"ContainerStarted","Data":"c7928a2bceb99f4b4564518463fb197c452d1977bbe5be608c6738681ef865ed"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.594935 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" event={"ID":"211eeae6-9b41-484b-bd13-99c1c28cdf96","Type":"ContainerStarted","Data":"b09c4ecfbce0d77ab595c1787ce759be04f8ca207eafee9fec5fb1a2e4086727"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.594946 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-n7pd7" event={"ID":"211eeae6-9b41-484b-bd13-99c1c28cdf96","Type":"ContainerStarted","Data":"fa523477acd29a14cfd122e3283db6ab6e928eb05746f29ea2b64f89366a97aa"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.597181 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" event={"ID":"23f96e9f-9853-4f3b-8c84-3ba79491d133","Type":"ContainerStarted","Data":"232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.597205 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" event={"ID":"23f96e9f-9853-4f3b-8c84-3ba79491d133","Type":"ContainerStarted","Data":"b7ab31a5c8fe6aa803d3d10a661aa331fd52cf721b1a3dc48f874195bb4a0809"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.597282 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" podUID="23f96e9f-9853-4f3b-8c84-3ba79491d133" containerName="controller-manager" containerID="cri-o://232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296" gracePeriod=30 Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.597678 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.601212 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" event={"ID":"91f58aee-72fa-449f-8ec9-c80f89e5bc39","Type":"ContainerStarted","Data":"a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.601239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" event={"ID":"91f58aee-72fa-449f-8ec9-c80f89e5bc39","Type":"ContainerStarted","Data":"3adfd4583ba63f15a8883e043574aecc69b295fca7e18080bc6085933f638518"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.601322 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" podUID="91f58aee-72fa-449f-8ec9-c80f89e5bc39" containerName="route-controller-manager" containerID="cri-o://a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66" gracePeriod=30 Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.601586 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.602484 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.609989 4926 patch_prober.go:28] interesting pod/route-controller-manager-b6dbb7676-l48qx container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": read tcp 10.217.0.2:38340->10.217.0.57:8443: read: connection reset by peer" start-of-body= Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.610035 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" podUID="91f58aee-72fa-449f-8ec9-c80f89e5bc39" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": read tcp 10.217.0.2:38340->10.217.0.57:8443: read: connection reset by peer" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.610303 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51ec2202-9e22-4586-af8c-020d00160c26","Type":"ContainerStarted","Data":"bd5240b017915f06ca1867df3058775494e0b616b22b2960334e2eca5d692bcd"} Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.610332 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51ec2202-9e22-4586-af8c-020d00160c26","Type":"ContainerStarted","Data":"7460ec72f709ffd09a704af8c7bdb191459d3a1e2566af183b451949d0331dba"} Mar 12 18:07:10 crc kubenswrapper[4926]: E0312 18:07:10.613745 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zx9h7" podUID="637236a6-6287-401d-a2cd-78713aa03176" Mar 12 18:07:10 crc kubenswrapper[4926]: E0312 18:07:10.615399 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zpnq8" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" Mar 12 18:07:10 crc kubenswrapper[4926]: E0312 18:07:10.615416 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-565fl" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" Mar 12 18:07:10 crc kubenswrapper[4926]: E0312 18:07:10.615485 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wz6qt" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.627865 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-n7pd7" podStartSLOduration=199.627845898 podStartE2EDuration="3m19.627845898s" podCreationTimestamp="2026-03-12 18:03:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:10.625105612 +0000 UTC m=+270.993731955" watchObservedRunningTime="2026-03-12 18:07:10.627845898 +0000 UTC m=+270.996472221" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.699371 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" podStartSLOduration=23.699354864 podStartE2EDuration="23.699354864s" podCreationTimestamp="2026-03-12 18:06:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:10.696142412 +0000 UTC m=+271.064768745" watchObservedRunningTime="2026-03-12 18:07:10.699354864 +0000 UTC m=+271.067981187" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.732123 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=4.732090766 podStartE2EDuration="4.732090766s" podCreationTimestamp="2026-03-12 18:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:10.729786044 +0000 UTC m=+271.098412387" watchObservedRunningTime="2026-03-12 18:07:10.732090766 +0000 UTC m=+271.100717099" Mar 12 18:07:10 crc kubenswrapper[4926]: I0312 18:07:10.778231 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" podStartSLOduration=23.778212172 podStartE2EDuration="23.778212172s" podCreationTimestamp="2026-03-12 18:06:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:10.776762926 +0000 UTC m=+271.145389279" watchObservedRunningTime="2026-03-12 18:07:10.778212172 +0000 UTC m=+271.146838505" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.072609 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.085271 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.103153 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d"] Mar 12 18:07:11 crc kubenswrapper[4926]: E0312 18:07:11.103409 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f58aee-72fa-449f-8ec9-c80f89e5bc39" containerName="route-controller-manager" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.103424 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f58aee-72fa-449f-8ec9-c80f89e5bc39" containerName="route-controller-manager" Mar 12 18:07:11 crc kubenswrapper[4926]: E0312 18:07:11.105606 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f96e9f-9853-4f3b-8c84-3ba79491d133" containerName="controller-manager" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.105663 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f96e9f-9853-4f3b-8c84-3ba79491d133" containerName="controller-manager" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.105933 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f58aee-72fa-449f-8ec9-c80f89e5bc39" containerName="route-controller-manager" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.105951 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="23f96e9f-9853-4f3b-8c84-3ba79491d133" containerName="controller-manager" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.106457 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.113010 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d"] Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.211752 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-client-ca\") pod \"23f96e9f-9853-4f3b-8c84-3ba79491d133\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.211819 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-proxy-ca-bundles\") pod \"23f96e9f-9853-4f3b-8c84-3ba79491d133\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.211864 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91f58aee-72fa-449f-8ec9-c80f89e5bc39-serving-cert\") pod \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.211902 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-config\") pod \"23f96e9f-9853-4f3b-8c84-3ba79491d133\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.211931 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxl78\" (UniqueName: \"kubernetes.io/projected/91f58aee-72fa-449f-8ec9-c80f89e5bc39-kube-api-access-jxl78\") pod \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.211982 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49sdm\" (UniqueName: \"kubernetes.io/projected/23f96e9f-9853-4f3b-8c84-3ba79491d133-kube-api-access-49sdm\") pod \"23f96e9f-9853-4f3b-8c84-3ba79491d133\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212031 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-config\") pod \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212066 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-client-ca\") pod \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\" (UID: \"91f58aee-72fa-449f-8ec9-c80f89e5bc39\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212098 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f96e9f-9853-4f3b-8c84-3ba79491d133-serving-cert\") pod \"23f96e9f-9853-4f3b-8c84-3ba79491d133\" (UID: \"23f96e9f-9853-4f3b-8c84-3ba79491d133\") " Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzg9f\" (UniqueName: \"kubernetes.io/projected/d9a46d07-aea8-4496-b088-1d30771449a5-kube-api-access-pzg9f\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212347 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-client-ca\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212385 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-proxy-ca-bundles\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212409 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-config\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212524 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-client-ca" (OuterVolumeSpecName: "client-ca") pod "23f96e9f-9853-4f3b-8c84-3ba79491d133" (UID: "23f96e9f-9853-4f3b-8c84-3ba79491d133"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.212582 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "23f96e9f-9853-4f3b-8c84-3ba79491d133" (UID: "23f96e9f-9853-4f3b-8c84-3ba79491d133"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.213338 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-config" (OuterVolumeSpecName: "config") pod "23f96e9f-9853-4f3b-8c84-3ba79491d133" (UID: "23f96e9f-9853-4f3b-8c84-3ba79491d133"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.213579 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-config" (OuterVolumeSpecName: "config") pod "91f58aee-72fa-449f-8ec9-c80f89e5bc39" (UID: "91f58aee-72fa-449f-8ec9-c80f89e5bc39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.213746 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a46d07-aea8-4496-b088-1d30771449a5-serving-cert\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.213877 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.213894 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.213908 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.213921 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f96e9f-9853-4f3b-8c84-3ba79491d133-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.214000 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-client-ca" (OuterVolumeSpecName: "client-ca") pod "91f58aee-72fa-449f-8ec9-c80f89e5bc39" (UID: "91f58aee-72fa-449f-8ec9-c80f89e5bc39"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.217763 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23f96e9f-9853-4f3b-8c84-3ba79491d133-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "23f96e9f-9853-4f3b-8c84-3ba79491d133" (UID: "23f96e9f-9853-4f3b-8c84-3ba79491d133"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.217934 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f58aee-72fa-449f-8ec9-c80f89e5bc39-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "91f58aee-72fa-449f-8ec9-c80f89e5bc39" (UID: "91f58aee-72fa-449f-8ec9-c80f89e5bc39"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.218227 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23f96e9f-9853-4f3b-8c84-3ba79491d133-kube-api-access-49sdm" (OuterVolumeSpecName: "kube-api-access-49sdm") pod "23f96e9f-9853-4f3b-8c84-3ba79491d133" (UID: "23f96e9f-9853-4f3b-8c84-3ba79491d133"). InnerVolumeSpecName "kube-api-access-49sdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.218306 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f58aee-72fa-449f-8ec9-c80f89e5bc39-kube-api-access-jxl78" (OuterVolumeSpecName: "kube-api-access-jxl78") pod "91f58aee-72fa-449f-8ec9-c80f89e5bc39" (UID: "91f58aee-72fa-449f-8ec9-c80f89e5bc39"). InnerVolumeSpecName "kube-api-access-jxl78". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.314909 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-config\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.314981 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a46d07-aea8-4496-b088-1d30771449a5-serving-cert\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315027 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzg9f\" (UniqueName: \"kubernetes.io/projected/d9a46d07-aea8-4496-b088-1d30771449a5-kube-api-access-pzg9f\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315057 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-client-ca\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315090 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-proxy-ca-bundles\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315131 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/91f58aee-72fa-449f-8ec9-c80f89e5bc39-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315143 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f96e9f-9853-4f3b-8c84-3ba79491d133-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315152 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91f58aee-72fa-449f-8ec9-c80f89e5bc39-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315162 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxl78\" (UniqueName: \"kubernetes.io/projected/91f58aee-72fa-449f-8ec9-c80f89e5bc39-kube-api-access-jxl78\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.315173 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49sdm\" (UniqueName: \"kubernetes.io/projected/23f96e9f-9853-4f3b-8c84-3ba79491d133-kube-api-access-49sdm\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.316263 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-client-ca\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.317501 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-config\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.319419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a46d07-aea8-4496-b088-1d30771449a5-serving-cert\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.319491 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-proxy-ca-bundles\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.330038 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzg9f\" (UniqueName: \"kubernetes.io/projected/d9a46d07-aea8-4496-b088-1d30771449a5-kube-api-access-pzg9f\") pod \"controller-manager-57d6b6c74b-ndw9d\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.480884 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.620804 4926 generic.go:334] "Generic (PLEG): container finished" podID="51ec2202-9e22-4586-af8c-020d00160c26" containerID="bd5240b017915f06ca1867df3058775494e0b616b22b2960334e2eca5d692bcd" exitCode=0 Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.620935 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51ec2202-9e22-4586-af8c-020d00160c26","Type":"ContainerDied","Data":"bd5240b017915f06ca1867df3058775494e0b616b22b2960334e2eca5d692bcd"} Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.624050 4926 generic.go:334] "Generic (PLEG): container finished" podID="91f58aee-72fa-449f-8ec9-c80f89e5bc39" containerID="a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66" exitCode=0 Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.624188 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.624242 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" event={"ID":"91f58aee-72fa-449f-8ec9-c80f89e5bc39","Type":"ContainerDied","Data":"a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66"} Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.624283 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx" event={"ID":"91f58aee-72fa-449f-8ec9-c80f89e5bc39","Type":"ContainerDied","Data":"3adfd4583ba63f15a8883e043574aecc69b295fca7e18080bc6085933f638518"} Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.624303 4926 scope.go:117] "RemoveContainer" containerID="a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.626828 4926 generic.go:334] "Generic (PLEG): container finished" podID="23f96e9f-9853-4f3b-8c84-3ba79491d133" containerID="232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296" exitCode=0 Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.627422 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.632574 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" event={"ID":"23f96e9f-9853-4f3b-8c84-3ba79491d133","Type":"ContainerDied","Data":"232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296"} Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.632626 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c9659bb69-k2bhk" event={"ID":"23f96e9f-9853-4f3b-8c84-3ba79491d133","Type":"ContainerDied","Data":"b7ab31a5c8fe6aa803d3d10a661aa331fd52cf721b1a3dc48f874195bb4a0809"} Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.654478 4926 scope.go:117] "RemoveContainer" containerID="a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66" Mar 12 18:07:11 crc kubenswrapper[4926]: E0312 18:07:11.655389 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66\": container with ID starting with a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66 not found: ID does not exist" containerID="a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.655485 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66"} err="failed to get container status \"a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66\": rpc error: code = NotFound desc = could not find container \"a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66\": container with ID starting with a14fd899c389a50963ad79b00b1c3a6ea8e5844f2a8769831d6f4a4c8d367d66 not found: ID does not exist" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.655516 4926 scope.go:117] "RemoveContainer" containerID="232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.663041 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c9659bb69-k2bhk"] Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.668478 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c9659bb69-k2bhk"] Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.671086 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx"] Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.673861 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-b6dbb7676-l48qx"] Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.677182 4926 scope.go:117] "RemoveContainer" containerID="232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296" Mar 12 18:07:11 crc kubenswrapper[4926]: E0312 18:07:11.677617 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296\": container with ID starting with 232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296 not found: ID does not exist" containerID="232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.677653 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296"} err="failed to get container status \"232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296\": rpc error: code = NotFound desc = could not find container \"232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296\": container with ID starting with 232bc05bcf31de44ab262bfed931d35f35e41de12c12d164f104762a81f8c296 not found: ID does not exist" Mar 12 18:07:11 crc kubenswrapper[4926]: I0312 18:07:11.859111 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d"] Mar 12 18:07:11 crc kubenswrapper[4926]: W0312 18:07:11.865078 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9a46d07_aea8_4496_b088_1d30771449a5.slice/crio-970f69b22a7b41275946a8af481797c27d777726022fcf08385e195d2cc5696f WatchSource:0}: Error finding container 970f69b22a7b41275946a8af481797c27d777726022fcf08385e195d2cc5696f: Status 404 returned error can't find the container with id 970f69b22a7b41275946a8af481797c27d777726022fcf08385e195d2cc5696f Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.442024 4926 csr.go:261] certificate signing request csr-kmkcd is approved, waiting to be issued Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.447909 4926 csr.go:257] certificate signing request csr-kmkcd is issued Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.503284 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23f96e9f-9853-4f3b-8c84-3ba79491d133" path="/var/lib/kubelet/pods/23f96e9f-9853-4f3b-8c84-3ba79491d133/volumes" Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.503780 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f58aee-72fa-449f-8ec9-c80f89e5bc39" path="/var/lib/kubelet/pods/91f58aee-72fa-449f-8ec9-c80f89e5bc39/volumes" Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.634581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" event={"ID":"d9a46d07-aea8-4496-b088-1d30771449a5","Type":"ContainerStarted","Data":"2e38aace17ed1402804edbaeeb451a49615dca3e7706295cc9da9a75fd5336e5"} Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.634852 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.634866 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" event={"ID":"d9a46d07-aea8-4496-b088-1d30771449a5","Type":"ContainerStarted","Data":"970f69b22a7b41275946a8af481797c27d777726022fcf08385e195d2cc5696f"} Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.636921 4926 generic.go:334] "Generic (PLEG): container finished" podID="d68160cf-4e6c-4294-bfdc-4acb74637ecb" containerID="52e8bc12e7903e1d736ebf550d285fa84f2e22f28d746c1ea7750000e16f3903" exitCode=0 Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.637239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" event={"ID":"d68160cf-4e6c-4294-bfdc-4acb74637ecb","Type":"ContainerDied","Data":"52e8bc12e7903e1d736ebf550d285fa84f2e22f28d746c1ea7750000e16f3903"} Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.641685 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.659713 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" podStartSLOduration=5.6596979229999995 podStartE2EDuration="5.659697923s" podCreationTimestamp="2026-03-12 18:07:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:12.655038086 +0000 UTC m=+273.023664419" watchObservedRunningTime="2026-03-12 18:07:12.659697923 +0000 UTC m=+273.028324256" Mar 12 18:07:12 crc kubenswrapper[4926]: I0312 18:07:12.920933 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.034747 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51ec2202-9e22-4586-af8c-020d00160c26-kube-api-access\") pod \"51ec2202-9e22-4586-af8c-020d00160c26\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.034847 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51ec2202-9e22-4586-af8c-020d00160c26-kubelet-dir\") pod \"51ec2202-9e22-4586-af8c-020d00160c26\" (UID: \"51ec2202-9e22-4586-af8c-020d00160c26\") " Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.034951 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51ec2202-9e22-4586-af8c-020d00160c26-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "51ec2202-9e22-4586-af8c-020d00160c26" (UID: "51ec2202-9e22-4586-af8c-020d00160c26"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.035129 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51ec2202-9e22-4586-af8c-020d00160c26-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.040047 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ec2202-9e22-4586-af8c-020d00160c26-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "51ec2202-9e22-4586-af8c-020d00160c26" (UID: "51ec2202-9e22-4586-af8c-020d00160c26"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.136892 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51ec2202-9e22-4586-af8c-020d00160c26-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.448944 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-05 13:33:33.890548427 +0000 UTC Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.448984 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7171h26m20.441566951s for next certificate rotation Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.644357 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51ec2202-9e22-4586-af8c-020d00160c26","Type":"ContainerDied","Data":"7460ec72f709ffd09a704af8c7bdb191459d3a1e2566af183b451949d0331dba"} Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.644504 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.644529 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7460ec72f709ffd09a704af8c7bdb191459d3a1e2566af183b451949d0331dba" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.780838 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq"] Mar 12 18:07:13 crc kubenswrapper[4926]: E0312 18:07:13.781164 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ec2202-9e22-4586-af8c-020d00160c26" containerName="pruner" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.781176 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ec2202-9e22-4586-af8c-020d00160c26" containerName="pruner" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.781275 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ec2202-9e22-4586-af8c-020d00160c26" containerName="pruner" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.781598 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq"] Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.781745 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.833637 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.833722 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.833859 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.833990 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.834006 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.834096 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.854365 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfl44\" (UniqueName: \"kubernetes.io/projected/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-kube-api-access-cfl44\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.854463 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-config\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.854497 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-client-ca\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.854841 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-serving-cert\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.945240 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.956208 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-serving-cert\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.956274 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfl44\" (UniqueName: \"kubernetes.io/projected/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-kube-api-access-cfl44\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.956308 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-config\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.956328 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-client-ca\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.958194 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-client-ca\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.958612 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-config\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.960517 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-serving-cert\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:13 crc kubenswrapper[4926]: I0312 18:07:13.972142 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfl44\" (UniqueName: \"kubernetes.io/projected/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-kube-api-access-cfl44\") pod \"route-controller-manager-56fbddc998-ghqvq\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.057183 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl7pp\" (UniqueName: \"kubernetes.io/projected/d68160cf-4e6c-4294-bfdc-4acb74637ecb-kube-api-access-tl7pp\") pod \"d68160cf-4e6c-4294-bfdc-4acb74637ecb\" (UID: \"d68160cf-4e6c-4294-bfdc-4acb74637ecb\") " Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.060220 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d68160cf-4e6c-4294-bfdc-4acb74637ecb-kube-api-access-tl7pp" (OuterVolumeSpecName: "kube-api-access-tl7pp") pod "d68160cf-4e6c-4294-bfdc-4acb74637ecb" (UID: "d68160cf-4e6c-4294-bfdc-4acb74637ecb"). InnerVolumeSpecName "kube-api-access-tl7pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.153072 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.158370 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl7pp\" (UniqueName: \"kubernetes.io/projected/d68160cf-4e6c-4294-bfdc-4acb74637ecb-kube-api-access-tl7pp\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.449523 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-13 11:26:27.293792942 +0000 UTC Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.449559 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5897h19m12.844235889s for next certificate rotation Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.534934 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.543713 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq"] Mar 12 18:07:14 crc kubenswrapper[4926]: W0312 18:07:14.557713 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ffc00d2_8ad7_4109_aab6_18c87d0bc51e.slice/crio-d9af210d4e051b95f3fb2613a9ee5e48096656a2e0c5a0f8ba2b6c022529ebcc WatchSource:0}: Error finding container d9af210d4e051b95f3fb2613a9ee5e48096656a2e0c5a0f8ba2b6c022529ebcc: Status 404 returned error can't find the container with id d9af210d4e051b95f3fb2613a9ee5e48096656a2e0c5a0f8ba2b6c022529ebcc Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.652260 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" event={"ID":"d68160cf-4e6c-4294-bfdc-4acb74637ecb","Type":"ContainerDied","Data":"71f803a395609858e2670b2b53c7204676afa8a634426e199a4c62b07d977042"} Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.652572 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71f803a395609858e2670b2b53c7204676afa8a634426e199a4c62b07d977042" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.652298 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555646-wqpkb" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.653591 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" event={"ID":"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e","Type":"ContainerStarted","Data":"d9af210d4e051b95f3fb2613a9ee5e48096656a2e0c5a0f8ba2b6c022529ebcc"} Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.825519 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 18:07:14 crc kubenswrapper[4926]: E0312 18:07:14.825786 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d68160cf-4e6c-4294-bfdc-4acb74637ecb" containerName="oc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.825806 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d68160cf-4e6c-4294-bfdc-4acb74637ecb" containerName="oc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.825926 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d68160cf-4e6c-4294-bfdc-4acb74637ecb" containerName="oc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.826341 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.828965 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.829162 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.832618 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.866398 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-var-lock\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.866512 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kube-api-access\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.866537 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.967674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-var-lock\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.967739 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kube-api-access\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.967754 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-var-lock\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.967772 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.967903 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kubelet-dir\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:14 crc kubenswrapper[4926]: I0312 18:07:14.988744 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kube-api-access\") pod \"installer-9-crc\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:15 crc kubenswrapper[4926]: I0312 18:07:15.151907 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:15 crc kubenswrapper[4926]: I0312 18:07:15.568404 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 18:07:15 crc kubenswrapper[4926]: W0312 18:07:15.572620 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod63c14bbd_6ba6_42c2_9e94_bf3a6f68500f.slice/crio-491a64d2c203ce09a50339566c321ae1dad16f74cdb93b33b26da5d1f82b4544 WatchSource:0}: Error finding container 491a64d2c203ce09a50339566c321ae1dad16f74cdb93b33b26da5d1f82b4544: Status 404 returned error can't find the container with id 491a64d2c203ce09a50339566c321ae1dad16f74cdb93b33b26da5d1f82b4544 Mar 12 18:07:15 crc kubenswrapper[4926]: I0312 18:07:15.660985 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f","Type":"ContainerStarted","Data":"491a64d2c203ce09a50339566c321ae1dad16f74cdb93b33b26da5d1f82b4544"} Mar 12 18:07:15 crc kubenswrapper[4926]: I0312 18:07:15.663147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" event={"ID":"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e","Type":"ContainerStarted","Data":"d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153"} Mar 12 18:07:15 crc kubenswrapper[4926]: I0312 18:07:15.663711 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:15 crc kubenswrapper[4926]: I0312 18:07:15.669512 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:15 crc kubenswrapper[4926]: I0312 18:07:15.682677 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" podStartSLOduration=8.682632268 podStartE2EDuration="8.682632268s" podCreationTimestamp="2026-03-12 18:07:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:15.679149188 +0000 UTC m=+276.047775531" watchObservedRunningTime="2026-03-12 18:07:15.682632268 +0000 UTC m=+276.051258631" Mar 12 18:07:16 crc kubenswrapper[4926]: I0312 18:07:16.669635 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f","Type":"ContainerStarted","Data":"988893c573b728d62d468c803dd6aa2333147e0f68cf77f807942af0ae0d7912"} Mar 12 18:07:16 crc kubenswrapper[4926]: I0312 18:07:16.684009 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.683997301 podStartE2EDuration="2.683997301s" podCreationTimestamp="2026-03-12 18:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:16.683565098 +0000 UTC m=+277.052191441" watchObservedRunningTime="2026-03-12 18:07:16.683997301 +0000 UTC m=+277.052623634" Mar 12 18:07:19 crc kubenswrapper[4926]: I0312 18:07:19.684620 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t4sbh" event={"ID":"150781c8-5ae3-42a6-b351-2388dfe84167","Type":"ContainerStarted","Data":"e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a"} Mar 12 18:07:19 crc kubenswrapper[4926]: I0312 18:07:19.686548 4926 generic.go:334] "Generic (PLEG): container finished" podID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerID="f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc" exitCode=0 Mar 12 18:07:19 crc kubenswrapper[4926]: I0312 18:07:19.686576 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4zf5" event={"ID":"b2a609cd-c298-4356-9ddf-a7f125b52938","Type":"ContainerDied","Data":"f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc"} Mar 12 18:07:20 crc kubenswrapper[4926]: I0312 18:07:20.693826 4926 generic.go:334] "Generic (PLEG): container finished" podID="150781c8-5ae3-42a6-b351-2388dfe84167" containerID="e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a" exitCode=0 Mar 12 18:07:20 crc kubenswrapper[4926]: I0312 18:07:20.693932 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t4sbh" event={"ID":"150781c8-5ae3-42a6-b351-2388dfe84167","Type":"ContainerDied","Data":"e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a"} Mar 12 18:07:20 crc kubenswrapper[4926]: I0312 18:07:20.701848 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4zf5" event={"ID":"b2a609cd-c298-4356-9ddf-a7f125b52938","Type":"ContainerStarted","Data":"1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e"} Mar 12 18:07:20 crc kubenswrapper[4926]: I0312 18:07:20.733298 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s4zf5" podStartSLOduration=3.648730875 podStartE2EDuration="50.733283957s" podCreationTimestamp="2026-03-12 18:06:30 +0000 UTC" firstStartedPulling="2026-03-12 18:06:33.018194031 +0000 UTC m=+233.386820364" lastFinishedPulling="2026-03-12 18:07:20.102747103 +0000 UTC m=+280.471373446" observedRunningTime="2026-03-12 18:07:20.73020616 +0000 UTC m=+281.098832493" watchObservedRunningTime="2026-03-12 18:07:20.733283957 +0000 UTC m=+281.101910290" Mar 12 18:07:21 crc kubenswrapper[4926]: I0312 18:07:21.417596 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:07:21 crc kubenswrapper[4926]: I0312 18:07:21.417655 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:07:21 crc kubenswrapper[4926]: I0312 18:07:21.712806 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t4sbh" event={"ID":"150781c8-5ae3-42a6-b351-2388dfe84167","Type":"ContainerStarted","Data":"2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970"} Mar 12 18:07:21 crc kubenswrapper[4926]: I0312 18:07:21.733113 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t4sbh" podStartSLOduration=2.363138664 podStartE2EDuration="50.733096272s" podCreationTimestamp="2026-03-12 18:06:31 +0000 UTC" firstStartedPulling="2026-03-12 18:06:33.023748026 +0000 UTC m=+233.392374359" lastFinishedPulling="2026-03-12 18:07:21.393705604 +0000 UTC m=+281.762331967" observedRunningTime="2026-03-12 18:07:21.727594998 +0000 UTC m=+282.096221351" watchObservedRunningTime="2026-03-12 18:07:21.733096272 +0000 UTC m=+282.101722615" Mar 12 18:07:22 crc kubenswrapper[4926]: I0312 18:07:22.561352 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-s4zf5" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="registry-server" probeResult="failure" output=< Mar 12 18:07:22 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:07:22 crc kubenswrapper[4926]: > Mar 12 18:07:23 crc kubenswrapper[4926]: I0312 18:07:23.727367 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerID="d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5" exitCode=0 Mar 12 18:07:23 crc kubenswrapper[4926]: I0312 18:07:23.727416 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-565fl" event={"ID":"b5fe4032-6a1e-4c27-9471-fa53e044826e","Type":"ContainerDied","Data":"d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5"} Mar 12 18:07:24 crc kubenswrapper[4926]: I0312 18:07:24.737238 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-565fl" event={"ID":"b5fe4032-6a1e-4c27-9471-fa53e044826e","Type":"ContainerStarted","Data":"a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8"} Mar 12 18:07:24 crc kubenswrapper[4926]: I0312 18:07:24.766722 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-565fl" podStartSLOduration=2.399645696 podStartE2EDuration="53.766704542s" podCreationTimestamp="2026-03-12 18:06:31 +0000 UTC" firstStartedPulling="2026-03-12 18:06:33.007413821 +0000 UTC m=+233.376040154" lastFinishedPulling="2026-03-12 18:07:24.374472627 +0000 UTC m=+284.743099000" observedRunningTime="2026-03-12 18:07:24.765090542 +0000 UTC m=+285.133716895" watchObservedRunningTime="2026-03-12 18:07:24.766704542 +0000 UTC m=+285.135330875" Mar 12 18:07:25 crc kubenswrapper[4926]: I0312 18:07:25.750877 4926 generic.go:334] "Generic (PLEG): container finished" podID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerID="23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0" exitCode=0 Mar 12 18:07:25 crc kubenswrapper[4926]: I0312 18:07:25.751003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5qw8" event={"ID":"8630848f-c268-4f4a-9fd0-8f33765c20b4","Type":"ContainerDied","Data":"23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0"} Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.761378 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5qw8" event={"ID":"8630848f-c268-4f4a-9fd0-8f33765c20b4","Type":"ContainerStarted","Data":"bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9"} Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.768838 4926 generic.go:334] "Generic (PLEG): container finished" podID="637236a6-6287-401d-a2cd-78713aa03176" containerID="599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb" exitCode=0 Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.768881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx9h7" event={"ID":"637236a6-6287-401d-a2cd-78713aa03176","Type":"ContainerDied","Data":"599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb"} Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.773712 4926 generic.go:334] "Generic (PLEG): container finished" podID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerID="ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199" exitCode=0 Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.773815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpnq8" event={"ID":"743a7318-33d0-4a59-93bd-7c6899554e5e","Type":"ContainerDied","Data":"ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199"} Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.777202 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmn6x" event={"ID":"1f425571-9ce5-4fdc-9631-7683efa292aa","Type":"ContainerStarted","Data":"9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429"} Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.779736 4926 generic.go:334] "Generic (PLEG): container finished" podID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerID="946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b" exitCode=0 Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.779778 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz6qt" event={"ID":"702daa2d-851e-4c3d-be86-4f337b4462f7","Type":"ContainerDied","Data":"946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b"} Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.785635 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v5qw8" podStartSLOduration=2.683320053 podStartE2EDuration="53.78561732s" podCreationTimestamp="2026-03-12 18:06:33 +0000 UTC" firstStartedPulling="2026-03-12 18:06:35.158601298 +0000 UTC m=+235.527227631" lastFinishedPulling="2026-03-12 18:07:26.260898555 +0000 UTC m=+286.629524898" observedRunningTime="2026-03-12 18:07:26.782552733 +0000 UTC m=+287.151179086" watchObservedRunningTime="2026-03-12 18:07:26.78561732 +0000 UTC m=+287.154243663" Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.817463 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.817540 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.817596 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.818271 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:07:26 crc kubenswrapper[4926]: I0312 18:07:26.818338 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff" gracePeriod=600 Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.161753 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d"] Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.162334 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" podUID="d9a46d07-aea8-4496-b088-1d30771449a5" containerName="controller-manager" containerID="cri-o://2e38aace17ed1402804edbaeeb451a49615dca3e7706295cc9da9a75fd5336e5" gracePeriod=30 Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.186398 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq"] Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.186671 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" podUID="5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" containerName="route-controller-manager" containerID="cri-o://d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153" gracePeriod=30 Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.686147 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.787021 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz6qt" event={"ID":"702daa2d-851e-4c3d-be86-4f337b4462f7","Type":"ContainerStarted","Data":"9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.788811 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff" exitCode=0 Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.788874 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.788923 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"b3aa39c92eb410de660dbb6e5cf8c0dd506addf3c227d2622f913bd6b55014e2"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.791119 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx9h7" event={"ID":"637236a6-6287-401d-a2cd-78713aa03176","Type":"ContainerStarted","Data":"2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.792307 4926 generic.go:334] "Generic (PLEG): container finished" podID="d9a46d07-aea8-4496-b088-1d30771449a5" containerID="2e38aace17ed1402804edbaeeb451a49615dca3e7706295cc9da9a75fd5336e5" exitCode=0 Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.792385 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" event={"ID":"d9a46d07-aea8-4496-b088-1d30771449a5","Type":"ContainerDied","Data":"2e38aace17ed1402804edbaeeb451a49615dca3e7706295cc9da9a75fd5336e5"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.795801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpnq8" event={"ID":"743a7318-33d0-4a59-93bd-7c6899554e5e","Type":"ContainerStarted","Data":"75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.796642 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.799264 4926 generic.go:334] "Generic (PLEG): container finished" podID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerID="9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429" exitCode=0 Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.799298 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmn6x" event={"ID":"1f425571-9ce5-4fdc-9631-7683efa292aa","Type":"ContainerDied","Data":"9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.801533 4926 generic.go:334] "Generic (PLEG): container finished" podID="5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" containerID="d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153" exitCode=0 Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.801566 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" event={"ID":"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e","Type":"ContainerDied","Data":"d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.801592 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" event={"ID":"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e","Type":"ContainerDied","Data":"d9af210d4e051b95f3fb2613a9ee5e48096656a2e0c5a0f8ba2b6c022529ebcc"} Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.801608 4926 scope.go:117] "RemoveContainer" containerID="d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.801741 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.832907 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wz6qt" podStartSLOduration=2.7855645239999998 podStartE2EDuration="54.832889551s" podCreationTimestamp="2026-03-12 18:06:33 +0000 UTC" firstStartedPulling="2026-03-12 18:06:35.197604664 +0000 UTC m=+235.566230997" lastFinishedPulling="2026-03-12 18:07:27.244929691 +0000 UTC m=+287.613556024" observedRunningTime="2026-03-12 18:07:27.814616715 +0000 UTC m=+288.183243048" watchObservedRunningTime="2026-03-12 18:07:27.832889551 +0000 UTC m=+288.201515884" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.833154 4926 scope.go:117] "RemoveContainer" containerID="d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153" Mar 12 18:07:27 crc kubenswrapper[4926]: E0312 18:07:27.833690 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153\": container with ID starting with d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153 not found: ID does not exist" containerID="d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.833791 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153"} err="failed to get container status \"d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153\": rpc error: code = NotFound desc = could not find container \"d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153\": container with ID starting with d7bed4139ba4a473385e0f0c19ea5e22e1f71a1f309b05c1696dda7157f26153 not found: ID does not exist" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.839363 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-serving-cert\") pod \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.839522 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-config\") pod \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.839598 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-client-ca\") pod \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.839706 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfl44\" (UniqueName: \"kubernetes.io/projected/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-kube-api-access-cfl44\") pod \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\" (UID: \"5ffc00d2-8ad7-4109-aab6-18c87d0bc51e\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.840750 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-config" (OuterVolumeSpecName: "config") pod "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" (UID: "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.841538 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-client-ca" (OuterVolumeSpecName: "client-ca") pod "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" (UID: "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.846243 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-kube-api-access-cfl44" (OuterVolumeSpecName: "kube-api-access-cfl44") pod "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" (UID: "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e"). InnerVolumeSpecName "kube-api-access-cfl44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.849236 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" (UID: "5ffc00d2-8ad7-4109-aab6-18c87d0bc51e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.857493 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zpnq8" podStartSLOduration=2.437658851 podStartE2EDuration="56.857474667s" podCreationTimestamp="2026-03-12 18:06:31 +0000 UTC" firstStartedPulling="2026-03-12 18:06:33.04112893 +0000 UTC m=+233.409755263" lastFinishedPulling="2026-03-12 18:07:27.460944746 +0000 UTC m=+287.829571079" observedRunningTime="2026-03-12 18:07:27.847758751 +0000 UTC m=+288.216385084" watchObservedRunningTime="2026-03-12 18:07:27.857474667 +0000 UTC m=+288.226101000" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.903371 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zx9h7" podStartSLOduration=2.940827531 podStartE2EDuration="53.903355855s" podCreationTimestamp="2026-03-12 18:06:34 +0000 UTC" firstStartedPulling="2026-03-12 18:06:36.23210791 +0000 UTC m=+236.600734243" lastFinishedPulling="2026-03-12 18:07:27.194636234 +0000 UTC m=+287.563262567" observedRunningTime="2026-03-12 18:07:27.90163736 +0000 UTC m=+288.270263693" watchObservedRunningTime="2026-03-12 18:07:27.903355855 +0000 UTC m=+288.271982188" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941101 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a46d07-aea8-4496-b088-1d30771449a5-serving-cert\") pod \"d9a46d07-aea8-4496-b088-1d30771449a5\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941170 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-proxy-ca-bundles\") pod \"d9a46d07-aea8-4496-b088-1d30771449a5\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941210 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzg9f\" (UniqueName: \"kubernetes.io/projected/d9a46d07-aea8-4496-b088-1d30771449a5-kube-api-access-pzg9f\") pod \"d9a46d07-aea8-4496-b088-1d30771449a5\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941261 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-config\") pod \"d9a46d07-aea8-4496-b088-1d30771449a5\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941293 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-client-ca\") pod \"d9a46d07-aea8-4496-b088-1d30771449a5\" (UID: \"d9a46d07-aea8-4496-b088-1d30771449a5\") " Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941722 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941758 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfl44\" (UniqueName: \"kubernetes.io/projected/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-kube-api-access-cfl44\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941772 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941782 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941944 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d9a46d07-aea8-4496-b088-1d30771449a5" (UID: "d9a46d07-aea8-4496-b088-1d30771449a5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.941997 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-config" (OuterVolumeSpecName: "config") pod "d9a46d07-aea8-4496-b088-1d30771449a5" (UID: "d9a46d07-aea8-4496-b088-1d30771449a5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.942275 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-client-ca" (OuterVolumeSpecName: "client-ca") pod "d9a46d07-aea8-4496-b088-1d30771449a5" (UID: "d9a46d07-aea8-4496-b088-1d30771449a5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.944023 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9a46d07-aea8-4496-b088-1d30771449a5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d9a46d07-aea8-4496-b088-1d30771449a5" (UID: "d9a46d07-aea8-4496-b088-1d30771449a5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:27 crc kubenswrapper[4926]: I0312 18:07:27.944993 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9a46d07-aea8-4496-b088-1d30771449a5-kube-api-access-pzg9f" (OuterVolumeSpecName: "kube-api-access-pzg9f") pod "d9a46d07-aea8-4496-b088-1d30771449a5" (UID: "d9a46d07-aea8-4496-b088-1d30771449a5"). InnerVolumeSpecName "kube-api-access-pzg9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.042453 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9a46d07-aea8-4496-b088-1d30771449a5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.042763 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.042778 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzg9f\" (UniqueName: \"kubernetes.io/projected/d9a46d07-aea8-4496-b088-1d30771449a5-kube-api-access-pzg9f\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.042790 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.042799 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9a46d07-aea8-4496-b088-1d30771449a5-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.129680 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq"] Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.135255 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56fbddc998-ghqvq"] Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.502625 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" path="/var/lib/kubelet/pods/5ffc00d2-8ad7-4109-aab6-18c87d0bc51e/volumes" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.788146 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74548584bf-x85ch"] Mar 12 18:07:28 crc kubenswrapper[4926]: E0312 18:07:28.788739 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" containerName="route-controller-manager" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.788757 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" containerName="route-controller-manager" Mar 12 18:07:28 crc kubenswrapper[4926]: E0312 18:07:28.788780 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a46d07-aea8-4496-b088-1d30771449a5" containerName="controller-manager" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.788788 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a46d07-aea8-4496-b088-1d30771449a5" containerName="controller-manager" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.788908 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9a46d07-aea8-4496-b088-1d30771449a5" containerName="controller-manager" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.788928 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ffc00d2-8ad7-4109-aab6-18c87d0bc51e" containerName="route-controller-manager" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.789555 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.791847 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2"] Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.792670 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.797114 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.797314 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.802827 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.802848 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.802963 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.803099 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.804756 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74548584bf-x85ch"] Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.810011 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2"] Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.812877 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmn6x" event={"ID":"1f425571-9ce5-4fdc-9631-7683efa292aa","Type":"ContainerStarted","Data":"9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167"} Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.815782 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" event={"ID":"d9a46d07-aea8-4496-b088-1d30771449a5","Type":"ContainerDied","Data":"970f69b22a7b41275946a8af481797c27d777726022fcf08385e195d2cc5696f"} Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.815828 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.815834 4926 scope.go:117] "RemoveContainer" containerID="2e38aace17ed1402804edbaeeb451a49615dca3e7706295cc9da9a75fd5336e5" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.842493 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d"] Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.847348 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-57d6b6c74b-ndw9d"] Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.864041 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hmn6x" podStartSLOduration=2.833563277 podStartE2EDuration="54.864026544s" podCreationTimestamp="2026-03-12 18:06:34 +0000 UTC" firstStartedPulling="2026-03-12 18:06:36.245100261 +0000 UTC m=+236.613726594" lastFinishedPulling="2026-03-12 18:07:28.275563528 +0000 UTC m=+288.644189861" observedRunningTime="2026-03-12 18:07:28.860213894 +0000 UTC m=+289.228840227" watchObservedRunningTime="2026-03-12 18:07:28.864026544 +0000 UTC m=+289.232652877" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.952926 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-serving-cert\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.952999 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw8k8\" (UniqueName: \"kubernetes.io/projected/be3d67c9-0e01-48ab-8091-2e0bf103655d-kube-api-access-mw8k8\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.953029 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3d67c9-0e01-48ab-8091-2e0bf103655d-serving-cert\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.953049 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-config\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.953071 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-client-ca\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.953091 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgt8s\" (UniqueName: \"kubernetes.io/projected/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-kube-api-access-zgt8s\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.953115 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-proxy-ca-bundles\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.953149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-config\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:28 crc kubenswrapper[4926]: I0312 18:07:28.953167 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-client-ca\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054290 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-config\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054342 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-client-ca\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054367 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-serving-cert\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw8k8\" (UniqueName: \"kubernetes.io/projected/be3d67c9-0e01-48ab-8091-2e0bf103655d-kube-api-access-mw8k8\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054449 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3d67c9-0e01-48ab-8091-2e0bf103655d-serving-cert\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054473 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-config\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054502 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-client-ca\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054521 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgt8s\" (UniqueName: \"kubernetes.io/projected/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-kube-api-access-zgt8s\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.054553 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-proxy-ca-bundles\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.055687 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-proxy-ca-bundles\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.056890 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-config\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.057581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-client-ca\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.058864 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-client-ca\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.059285 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-config\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.077789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw8k8\" (UniqueName: \"kubernetes.io/projected/be3d67c9-0e01-48ab-8091-2e0bf103655d-kube-api-access-mw8k8\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.079255 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3d67c9-0e01-48ab-8091-2e0bf103655d-serving-cert\") pod \"route-controller-manager-6b5767577f-5znp2\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.080179 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgt8s\" (UniqueName: \"kubernetes.io/projected/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-kube-api-access-zgt8s\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.080659 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-serving-cert\") pod \"controller-manager-74548584bf-x85ch\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.106840 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.116131 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.319973 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74548584bf-x85ch"] Mar 12 18:07:29 crc kubenswrapper[4926]: W0312 18:07:29.324748 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5f25640_5527_4caa_a290_cbb3bbfc5e0b.slice/crio-9b5a165d0fdbcf22977eee75175ce438f8c21d2c260b57e0ebed1521f9e89fb5 WatchSource:0}: Error finding container 9b5a165d0fdbcf22977eee75175ce438f8c21d2c260b57e0ebed1521f9e89fb5: Status 404 returned error can't find the container with id 9b5a165d0fdbcf22977eee75175ce438f8c21d2c260b57e0ebed1521f9e89fb5 Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.659724 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2"] Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.816135 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" podUID="eedd886d-5443-47e1-afbf-5aff90067f3b" containerName="oauth-openshift" containerID="cri-o://9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98" gracePeriod=15 Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.856659 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" event={"ID":"a5f25640-5527-4caa-a290-cbb3bbfc5e0b","Type":"ContainerStarted","Data":"4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2"} Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.856927 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" event={"ID":"a5f25640-5527-4caa-a290-cbb3bbfc5e0b","Type":"ContainerStarted","Data":"9b5a165d0fdbcf22977eee75175ce438f8c21d2c260b57e0ebed1521f9e89fb5"} Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.857945 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.873806 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" event={"ID":"be3d67c9-0e01-48ab-8091-2e0bf103655d","Type":"ContainerStarted","Data":"19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128"} Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.874038 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" event={"ID":"be3d67c9-0e01-48ab-8091-2e0bf103655d","Type":"ContainerStarted","Data":"581ff73fc13e04cb81a4a9c9a97934c14adf3bf26fd03837864c6f60ecb97184"} Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.874930 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.878169 4926 patch_prober.go:28] interesting pod/route-controller-manager-6b5767577f-5znp2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.878219 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" podUID="be3d67c9-0e01-48ab-8091-2e0bf103655d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.905986 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.910540 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" podStartSLOduration=2.910526431 podStartE2EDuration="2.910526431s" podCreationTimestamp="2026-03-12 18:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:29.88672585 +0000 UTC m=+290.255352183" watchObservedRunningTime="2026-03-12 18:07:29.910526431 +0000 UTC m=+290.279152754" Mar 12 18:07:29 crc kubenswrapper[4926]: I0312 18:07:29.911758 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" podStartSLOduration=2.91175347 podStartE2EDuration="2.91175347s" podCreationTimestamp="2026-03-12 18:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:29.908780986 +0000 UTC m=+290.277407319" watchObservedRunningTime="2026-03-12 18:07:29.91175347 +0000 UTC m=+290.280379803" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.223481 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379468 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-dir\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-session\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379573 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379631 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-ocp-branding-template\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379694 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-policies\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379728 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-error\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379763 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-serving-cert\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379798 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-provider-selection\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379832 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-idp-0-file-data\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.379870 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88hg4\" (UniqueName: \"kubernetes.io/projected/eedd886d-5443-47e1-afbf-5aff90067f3b-kube-api-access-88hg4\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.380897 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-router-certs\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.380952 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-cliconfig\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.381005 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-service-ca\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.381056 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-login\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.381099 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-trusted-ca-bundle\") pod \"eedd886d-5443-47e1-afbf-5aff90067f3b\" (UID: \"eedd886d-5443-47e1-afbf-5aff90067f3b\") " Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.381353 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.381507 4926 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.381519 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.382154 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.382256 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.382404 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.405290 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.405726 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.406029 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.407957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.409106 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eedd886d-5443-47e1-afbf-5aff90067f3b-kube-api-access-88hg4" (OuterVolumeSpecName: "kube-api-access-88hg4") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "kube-api-access-88hg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.409202 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.409458 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.412718 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.414366 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "eedd886d-5443-47e1-afbf-5aff90067f3b" (UID: "eedd886d-5443-47e1-afbf-5aff90067f3b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482528 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482569 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482579 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482590 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482603 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88hg4\" (UniqueName: \"kubernetes.io/projected/eedd886d-5443-47e1-afbf-5aff90067f3b-kube-api-access-88hg4\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482614 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482623 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482634 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482643 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482654 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482664 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.482675 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eedd886d-5443-47e1-afbf-5aff90067f3b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.505540 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9a46d07-aea8-4496-b088-1d30771449a5" path="/var/lib/kubelet/pods/d9a46d07-aea8-4496-b088-1d30771449a5/volumes" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.880003 4926 generic.go:334] "Generic (PLEG): container finished" podID="eedd886d-5443-47e1-afbf-5aff90067f3b" containerID="9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98" exitCode=0 Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.880949 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.881038 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" event={"ID":"eedd886d-5443-47e1-afbf-5aff90067f3b","Type":"ContainerDied","Data":"9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98"} Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.881062 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7rmsn" event={"ID":"eedd886d-5443-47e1-afbf-5aff90067f3b","Type":"ContainerDied","Data":"f375d41ba80f860b9235fe8ebb0d083845e27083ae0a2e004437a84b6b1f824e"} Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.881079 4926 scope.go:117] "RemoveContainer" containerID="9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.885664 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.896887 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7rmsn"] Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.899689 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7rmsn"] Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.903745 4926 scope.go:117] "RemoveContainer" containerID="9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98" Mar 12 18:07:30 crc kubenswrapper[4926]: E0312 18:07:30.904511 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98\": container with ID starting with 9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98 not found: ID does not exist" containerID="9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98" Mar 12 18:07:30 crc kubenswrapper[4926]: I0312 18:07:30.904557 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98"} err="failed to get container status \"9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98\": rpc error: code = NotFound desc = could not find container \"9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98\": container with ID starting with 9eb303537b742a873d0e271421bd5a8464795ffe3ddf357bdfd78ea7386ecc98 not found: ID does not exist" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.479667 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.494870 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.494905 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.537010 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.541271 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.687238 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.687567 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.728971 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.881282 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.881352 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.932171 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.943207 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:07:31 crc kubenswrapper[4926]: I0312 18:07:31.944376 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:07:32 crc kubenswrapper[4926]: I0312 18:07:32.504489 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eedd886d-5443-47e1-afbf-5aff90067f3b" path="/var/lib/kubelet/pods/eedd886d-5443-47e1-afbf-5aff90067f3b/volumes" Mar 12 18:07:32 crc kubenswrapper[4926]: I0312 18:07:32.948086 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:07:33 crc kubenswrapper[4926]: I0312 18:07:33.555357 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:07:33 crc kubenswrapper[4926]: I0312 18:07:33.555428 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:07:33 crc kubenswrapper[4926]: I0312 18:07:33.619479 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:07:33 crc kubenswrapper[4926]: I0312 18:07:33.904941 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:07:33 crc kubenswrapper[4926]: I0312 18:07:33.904999 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:07:33 crc kubenswrapper[4926]: I0312 18:07:33.952772 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:07:33 crc kubenswrapper[4926]: I0312 18:07:33.965515 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:07:34 crc kubenswrapper[4926]: I0312 18:07:34.129269 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t4sbh"] Mar 12 18:07:34 crc kubenswrapper[4926]: I0312 18:07:34.504729 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:07:34 crc kubenswrapper[4926]: I0312 18:07:34.504998 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:07:34 crc kubenswrapper[4926]: I0312 18:07:34.896920 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:07:34 crc kubenswrapper[4926]: I0312 18:07:34.897049 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:07:34 crc kubenswrapper[4926]: I0312 18:07:34.913194 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t4sbh" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="registry-server" containerID="cri-o://2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970" gracePeriod=2 Mar 12 18:07:34 crc kubenswrapper[4926]: I0312 18:07:34.978933 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.351227 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.457893 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-utilities\") pod \"150781c8-5ae3-42a6-b351-2388dfe84167\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.457938 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnjm4\" (UniqueName: \"kubernetes.io/projected/150781c8-5ae3-42a6-b351-2388dfe84167-kube-api-access-tnjm4\") pod \"150781c8-5ae3-42a6-b351-2388dfe84167\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.457963 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-catalog-content\") pod \"150781c8-5ae3-42a6-b351-2388dfe84167\" (UID: \"150781c8-5ae3-42a6-b351-2388dfe84167\") " Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.459188 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-utilities" (OuterVolumeSpecName: "utilities") pod "150781c8-5ae3-42a6-b351-2388dfe84167" (UID: "150781c8-5ae3-42a6-b351-2388dfe84167"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.465693 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.467787 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/150781c8-5ae3-42a6-b351-2388dfe84167-kube-api-access-tnjm4" (OuterVolumeSpecName: "kube-api-access-tnjm4") pod "150781c8-5ae3-42a6-b351-2388dfe84167" (UID: "150781c8-5ae3-42a6-b351-2388dfe84167"). InnerVolumeSpecName "kube-api-access-tnjm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.512046 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "150781c8-5ae3-42a6-b351-2388dfe84167" (UID: "150781c8-5ae3-42a6-b351-2388dfe84167"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.550780 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zx9h7" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="registry-server" probeResult="failure" output=< Mar 12 18:07:35 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:07:35 crc kubenswrapper[4926]: > Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.567389 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnjm4\" (UniqueName: \"kubernetes.io/projected/150781c8-5ae3-42a6-b351-2388dfe84167-kube-api-access-tnjm4\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.567461 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/150781c8-5ae3-42a6-b351-2388dfe84167-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.920472 4926 generic.go:334] "Generic (PLEG): container finished" podID="150781c8-5ae3-42a6-b351-2388dfe84167" containerID="2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970" exitCode=0 Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.920721 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t4sbh" event={"ID":"150781c8-5ae3-42a6-b351-2388dfe84167","Type":"ContainerDied","Data":"2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970"} Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.920698 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t4sbh" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.921060 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t4sbh" event={"ID":"150781c8-5ae3-42a6-b351-2388dfe84167","Type":"ContainerDied","Data":"f0734678dc103280ac23f9e4a0964bfd2dcfda20bb5ae8365cb587d30b175b62"} Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.921077 4926 scope.go:117] "RemoveContainer" containerID="2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.926828 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpnq8"] Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.927072 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zpnq8" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="registry-server" containerID="cri-o://75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff" gracePeriod=2 Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.944586 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hmn6x" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="registry-server" probeResult="failure" output=< Mar 12 18:07:35 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:07:35 crc kubenswrapper[4926]: > Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.944747 4926 scope.go:117] "RemoveContainer" containerID="e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.955127 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t4sbh"] Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.958148 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t4sbh"] Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.967087 4926 scope.go:117] "RemoveContainer" containerID="2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.991279 4926 scope.go:117] "RemoveContainer" containerID="2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970" Mar 12 18:07:35 crc kubenswrapper[4926]: E0312 18:07:35.992699 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970\": container with ID starting with 2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970 not found: ID does not exist" containerID="2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.992734 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970"} err="failed to get container status \"2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970\": rpc error: code = NotFound desc = could not find container \"2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970\": container with ID starting with 2edc573e204b1cec4545170d5c4313584b0e49b6fa8db612c1be47281fb04970 not found: ID does not exist" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.992753 4926 scope.go:117] "RemoveContainer" containerID="e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a" Mar 12 18:07:35 crc kubenswrapper[4926]: E0312 18:07:35.993164 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a\": container with ID starting with e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a not found: ID does not exist" containerID="e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.993231 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a"} err="failed to get container status \"e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a\": rpc error: code = NotFound desc = could not find container \"e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a\": container with ID starting with e271f20c168fa4ef1ea9fdd75f78dc064672cebc67bc3259aae0ce8d79cdc29a not found: ID does not exist" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.993291 4926 scope.go:117] "RemoveContainer" containerID="2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139" Mar 12 18:07:35 crc kubenswrapper[4926]: E0312 18:07:35.993636 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139\": container with ID starting with 2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139 not found: ID does not exist" containerID="2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139" Mar 12 18:07:35 crc kubenswrapper[4926]: I0312 18:07:35.993676 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139"} err="failed to get container status \"2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139\": rpc error: code = NotFound desc = could not find container \"2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139\": container with ID starting with 2262d4d70a418b07a3dcb18723e596af882371729d1b0168b7d72456d55f1139 not found: ID does not exist" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.389064 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.478287 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-utilities\") pod \"743a7318-33d0-4a59-93bd-7c6899554e5e\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.478346 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsdsv\" (UniqueName: \"kubernetes.io/projected/743a7318-33d0-4a59-93bd-7c6899554e5e-kube-api-access-fsdsv\") pod \"743a7318-33d0-4a59-93bd-7c6899554e5e\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.478394 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-catalog-content\") pod \"743a7318-33d0-4a59-93bd-7c6899554e5e\" (UID: \"743a7318-33d0-4a59-93bd-7c6899554e5e\") " Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.480645 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-utilities" (OuterVolumeSpecName: "utilities") pod "743a7318-33d0-4a59-93bd-7c6899554e5e" (UID: "743a7318-33d0-4a59-93bd-7c6899554e5e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.481964 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.484815 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743a7318-33d0-4a59-93bd-7c6899554e5e-kube-api-access-fsdsv" (OuterVolumeSpecName: "kube-api-access-fsdsv") pod "743a7318-33d0-4a59-93bd-7c6899554e5e" (UID: "743a7318-33d0-4a59-93bd-7c6899554e5e"). InnerVolumeSpecName "kube-api-access-fsdsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.504149 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" path="/var/lib/kubelet/pods/150781c8-5ae3-42a6-b351-2388dfe84167/volumes" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.526043 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5qw8"] Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.534287 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "743a7318-33d0-4a59-93bd-7c6899554e5e" (UID: "743a7318-33d0-4a59-93bd-7c6899554e5e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.583818 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743a7318-33d0-4a59-93bd-7c6899554e5e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.583872 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsdsv\" (UniqueName: \"kubernetes.io/projected/743a7318-33d0-4a59-93bd-7c6899554e5e-kube-api-access-fsdsv\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.935511 4926 generic.go:334] "Generic (PLEG): container finished" podID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerID="75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff" exitCode=0 Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.935566 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpnq8" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.935612 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpnq8" event={"ID":"743a7318-33d0-4a59-93bd-7c6899554e5e","Type":"ContainerDied","Data":"75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff"} Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.935645 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpnq8" event={"ID":"743a7318-33d0-4a59-93bd-7c6899554e5e","Type":"ContainerDied","Data":"dc92306c9f92008ccb5e4b67f4e37a7020d96c316ccb9bc2ecf40d9ac23d9cf2"} Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.935666 4926 scope.go:117] "RemoveContainer" containerID="75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.935966 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v5qw8" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="registry-server" containerID="cri-o://bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9" gracePeriod=2 Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.960668 4926 scope.go:117] "RemoveContainer" containerID="ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199" Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.974213 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zpnq8"] Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.975394 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zpnq8"] Mar 12 18:07:36 crc kubenswrapper[4926]: I0312 18:07:36.983700 4926 scope.go:117] "RemoveContainer" containerID="80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.061768 4926 scope.go:117] "RemoveContainer" containerID="75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.062471 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff\": container with ID starting with 75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff not found: ID does not exist" containerID="75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.062502 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff"} err="failed to get container status \"75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff\": rpc error: code = NotFound desc = could not find container \"75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff\": container with ID starting with 75d3832859836653ff1e10c1d5dba06996ff45d8969ab164fd1416b7403fd4ff not found: ID does not exist" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.062530 4926 scope.go:117] "RemoveContainer" containerID="ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.062938 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199\": container with ID starting with ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199 not found: ID does not exist" containerID="ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.063027 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199"} err="failed to get container status \"ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199\": rpc error: code = NotFound desc = could not find container \"ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199\": container with ID starting with ebf537f8f44f9ee169e4b4c6b4fc14ae60e34fc19d7bb85795f2dd414478a199 not found: ID does not exist" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.063082 4926 scope.go:117] "RemoveContainer" containerID="80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.063824 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9\": container with ID starting with 80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9 not found: ID does not exist" containerID="80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.063851 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9"} err="failed to get container status \"80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9\": rpc error: code = NotFound desc = could not find container \"80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9\": container with ID starting with 80367076dd605a92b7bdc13d1f48d31cd8288852dc66b581eb496ee2193179d9 not found: ID does not exist" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.427106 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.492906 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-catalog-content\") pod \"8630848f-c268-4f4a-9fd0-8f33765c20b4\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.492977 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-utilities\") pod \"8630848f-c268-4f4a-9fd0-8f33765c20b4\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.493012 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkgrh\" (UniqueName: \"kubernetes.io/projected/8630848f-c268-4f4a-9fd0-8f33765c20b4-kube-api-access-pkgrh\") pod \"8630848f-c268-4f4a-9fd0-8f33765c20b4\" (UID: \"8630848f-c268-4f4a-9fd0-8f33765c20b4\") " Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.494421 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-utilities" (OuterVolumeSpecName: "utilities") pod "8630848f-c268-4f4a-9fd0-8f33765c20b4" (UID: "8630848f-c268-4f4a-9fd0-8f33765c20b4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.497740 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8630848f-c268-4f4a-9fd0-8f33765c20b4-kube-api-access-pkgrh" (OuterVolumeSpecName: "kube-api-access-pkgrh") pod "8630848f-c268-4f4a-9fd0-8f33765c20b4" (UID: "8630848f-c268-4f4a-9fd0-8f33765c20b4"). InnerVolumeSpecName "kube-api-access-pkgrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.521349 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8630848f-c268-4f4a-9fd0-8f33765c20b4" (UID: "8630848f-c268-4f4a-9fd0-8f33765c20b4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.594537 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.594571 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8630848f-c268-4f4a-9fd0-8f33765c20b4-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.594582 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkgrh\" (UniqueName: \"kubernetes.io/projected/8630848f-c268-4f4a-9fd0-8f33765c20b4-kube-api-access-pkgrh\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.800679 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-68755f559b-zw9n6"] Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.801802 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="extract-content" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.801851 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="extract-content" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.801900 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.801919 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.801996 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802015 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.802039 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedd886d-5443-47e1-afbf-5aff90067f3b" containerName="oauth-openshift" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802057 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedd886d-5443-47e1-afbf-5aff90067f3b" containerName="oauth-openshift" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.802113 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="extract-content" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802131 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="extract-content" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.802157 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="extract-content" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802174 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="extract-content" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.802212 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="extract-utilities" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802231 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="extract-utilities" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.802270 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="extract-utilities" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802288 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="extract-utilities" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.802353 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802373 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: E0312 18:07:37.802406 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="extract-utilities" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.802424 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="extract-utilities" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.803184 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.803235 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="150781c8-5ae3-42a6-b351-2388dfe84167" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.803278 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerName="registry-server" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.803302 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eedd886d-5443-47e1-afbf-5aff90067f3b" containerName="oauth-openshift" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.804593 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.863085 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.863129 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.863181 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.863219 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.863314 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.863867 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.865047 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.865513 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.869424 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.870931 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.872166 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.872202 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.879539 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.880639 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.889660 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68755f559b-zw9n6"] Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.896951 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.897933 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898004 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp2j4\" (UniqueName: \"kubernetes.io/projected/6504d75a-184b-4809-bdaf-13993983da43-kube-api-access-sp2j4\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898054 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-error\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898077 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898117 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898141 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898171 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6504d75a-184b-4809-bdaf-13993983da43-audit-dir\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898221 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-session\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898254 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-login\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898319 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898337 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898540 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-service-ca\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898575 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-router-certs\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.898613 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-audit-policies\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.948212 4926 generic.go:334] "Generic (PLEG): container finished" podID="8630848f-c268-4f4a-9fd0-8f33765c20b4" containerID="bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9" exitCode=0 Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.948261 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5qw8" event={"ID":"8630848f-c268-4f4a-9fd0-8f33765c20b4","Type":"ContainerDied","Data":"bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9"} Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.948289 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5qw8" event={"ID":"8630848f-c268-4f4a-9fd0-8f33765c20b4","Type":"ContainerDied","Data":"e7bb37ba860cfb3ecfc47e3421fdc163d8b676f30a5d56c59a6b61ad57e2cd0c"} Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.948313 4926 scope.go:117] "RemoveContainer" containerID="bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.948506 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5qw8" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.974910 4926 scope.go:117] "RemoveContainer" containerID="23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999265 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-service-ca\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999313 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-router-certs\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999341 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-audit-policies\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999349 4926 scope.go:117] "RemoveContainer" containerID="ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999386 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp2j4\" (UniqueName: \"kubernetes.io/projected/6504d75a-184b-4809-bdaf-13993983da43-kube-api-access-sp2j4\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999411 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-error\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999452 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999476 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999469 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5qw8"] Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999501 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999529 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6504d75a-184b-4809-bdaf-13993983da43-audit-dir\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999556 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-session\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999585 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-login\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999617 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:37 crc kubenswrapper[4926]: I0312 18:07:37.999637 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.000717 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-service-ca\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.000780 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-audit-policies\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.001001 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.001361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6504d75a-184b-4809-bdaf-13993983da43-audit-dir\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.002716 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.006787 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-error\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.006805 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.008989 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-router-certs\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.009023 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-template-login\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.010160 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5qw8"] Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.011052 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.012156 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-session\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.013256 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.015047 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6504d75a-184b-4809-bdaf-13993983da43-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.037089 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp2j4\" (UniqueName: \"kubernetes.io/projected/6504d75a-184b-4809-bdaf-13993983da43-kube-api-access-sp2j4\") pod \"oauth-openshift-68755f559b-zw9n6\" (UID: \"6504d75a-184b-4809-bdaf-13993983da43\") " pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.038978 4926 scope.go:117] "RemoveContainer" containerID="bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9" Mar 12 18:07:38 crc kubenswrapper[4926]: E0312 18:07:38.039323 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9\": container with ID starting with bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9 not found: ID does not exist" containerID="bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.039371 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9"} err="failed to get container status \"bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9\": rpc error: code = NotFound desc = could not find container \"bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9\": container with ID starting with bab8d0e3055afeafd4a0d6f6f1f368c0c90798b28e54a364754943c7018fc4a9 not found: ID does not exist" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.039402 4926 scope.go:117] "RemoveContainer" containerID="23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0" Mar 12 18:07:38 crc kubenswrapper[4926]: E0312 18:07:38.039936 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0\": container with ID starting with 23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0 not found: ID does not exist" containerID="23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.039969 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0"} err="failed to get container status \"23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0\": rpc error: code = NotFound desc = could not find container \"23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0\": container with ID starting with 23e4b185b3be270175429e9780ce6edd9c856427857e8483640952f1aea32ee0 not found: ID does not exist" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.039990 4926 scope.go:117] "RemoveContainer" containerID="ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3" Mar 12 18:07:38 crc kubenswrapper[4926]: E0312 18:07:38.040174 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3\": container with ID starting with ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3 not found: ID does not exist" containerID="ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.040204 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3"} err="failed to get container status \"ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3\": rpc error: code = NotFound desc = could not find container \"ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3\": container with ID starting with ed7f7979c7884d3e906be4f3ddd912adf35a0794bb60ca54cf1534b508fe81a3 not found: ID does not exist" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.186818 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.498381 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743a7318-33d0-4a59-93bd-7c6899554e5e" path="/var/lib/kubelet/pods/743a7318-33d0-4a59-93bd-7c6899554e5e/volumes" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.499070 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8630848f-c268-4f4a-9fd0-8f33765c20b4" path="/var/lib/kubelet/pods/8630848f-c268-4f4a-9fd0-8f33765c20b4/volumes" Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.632355 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68755f559b-zw9n6"] Mar 12 18:07:38 crc kubenswrapper[4926]: W0312 18:07:38.637485 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6504d75a_184b_4809_bdaf_13993983da43.slice/crio-5a3d486791cdcdc8b87a6bcc221dcab256a616945fc43dc009396dd8982ea0cd WatchSource:0}: Error finding container 5a3d486791cdcdc8b87a6bcc221dcab256a616945fc43dc009396dd8982ea0cd: Status 404 returned error can't find the container with id 5a3d486791cdcdc8b87a6bcc221dcab256a616945fc43dc009396dd8982ea0cd Mar 12 18:07:38 crc kubenswrapper[4926]: I0312 18:07:38.956272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" event={"ID":"6504d75a-184b-4809-bdaf-13993983da43","Type":"ContainerStarted","Data":"5a3d486791cdcdc8b87a6bcc221dcab256a616945fc43dc009396dd8982ea0cd"} Mar 12 18:07:39 crc kubenswrapper[4926]: I0312 18:07:39.966289 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" event={"ID":"6504d75a-184b-4809-bdaf-13993983da43","Type":"ContainerStarted","Data":"87287b9d71746c3893271027c2a3f96b1867e2e5354267f2c6fb5b64c9c65375"} Mar 12 18:07:39 crc kubenswrapper[4926]: I0312 18:07:39.966680 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:39 crc kubenswrapper[4926]: I0312 18:07:39.975901 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" Mar 12 18:07:39 crc kubenswrapper[4926]: I0312 18:07:39.995959 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-68755f559b-zw9n6" podStartSLOduration=35.995930905 podStartE2EDuration="35.995930905s" podCreationTimestamp="2026-03-12 18:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:39.992329082 +0000 UTC m=+300.360955445" watchObservedRunningTime="2026-03-12 18:07:39.995930905 +0000 UTC m=+300.364557268" Mar 12 18:07:44 crc kubenswrapper[4926]: I0312 18:07:44.554335 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:07:44 crc kubenswrapper[4926]: I0312 18:07:44.600659 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:07:44 crc kubenswrapper[4926]: I0312 18:07:44.940491 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:07:44 crc kubenswrapper[4926]: I0312 18:07:44.982667 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.163563 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74548584bf-x85ch"] Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.164084 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" podUID="a5f25640-5527-4caa-a290-cbb3bbfc5e0b" containerName="controller-manager" containerID="cri-o://4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2" gracePeriod=30 Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.248547 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2"] Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.248774 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" podUID="be3d67c9-0e01-48ab-8091-2e0bf103655d" containerName="route-controller-manager" containerID="cri-o://19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128" gracePeriod=30 Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.663622 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.670938 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725427 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-config\") pod \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725548 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-serving-cert\") pod \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725578 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgt8s\" (UniqueName: \"kubernetes.io/projected/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-kube-api-access-zgt8s\") pod \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725603 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-proxy-ca-bundles\") pod \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725629 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw8k8\" (UniqueName: \"kubernetes.io/projected/be3d67c9-0e01-48ab-8091-2e0bf103655d-kube-api-access-mw8k8\") pod \"be3d67c9-0e01-48ab-8091-2e0bf103655d\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725650 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-config\") pod \"be3d67c9-0e01-48ab-8091-2e0bf103655d\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725682 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-client-ca\") pod \"be3d67c9-0e01-48ab-8091-2e0bf103655d\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725697 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-client-ca\") pod \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\" (UID: \"a5f25640-5527-4caa-a290-cbb3bbfc5e0b\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.725713 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3d67c9-0e01-48ab-8091-2e0bf103655d-serving-cert\") pod \"be3d67c9-0e01-48ab-8091-2e0bf103655d\" (UID: \"be3d67c9-0e01-48ab-8091-2e0bf103655d\") " Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.726353 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a5f25640-5527-4caa-a290-cbb3bbfc5e0b" (UID: "a5f25640-5527-4caa-a290-cbb3bbfc5e0b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.726876 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-client-ca" (OuterVolumeSpecName: "client-ca") pod "a5f25640-5527-4caa-a290-cbb3bbfc5e0b" (UID: "a5f25640-5527-4caa-a290-cbb3bbfc5e0b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.727182 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-client-ca" (OuterVolumeSpecName: "client-ca") pod "be3d67c9-0e01-48ab-8091-2e0bf103655d" (UID: "be3d67c9-0e01-48ab-8091-2e0bf103655d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.727294 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-config" (OuterVolumeSpecName: "config") pod "be3d67c9-0e01-48ab-8091-2e0bf103655d" (UID: "be3d67c9-0e01-48ab-8091-2e0bf103655d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.727328 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-config" (OuterVolumeSpecName: "config") pod "a5f25640-5527-4caa-a290-cbb3bbfc5e0b" (UID: "a5f25640-5527-4caa-a290-cbb3bbfc5e0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.731133 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be3d67c9-0e01-48ab-8091-2e0bf103655d-kube-api-access-mw8k8" (OuterVolumeSpecName: "kube-api-access-mw8k8") pod "be3d67c9-0e01-48ab-8091-2e0bf103655d" (UID: "be3d67c9-0e01-48ab-8091-2e0bf103655d"). InnerVolumeSpecName "kube-api-access-mw8k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.731202 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be3d67c9-0e01-48ab-8091-2e0bf103655d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "be3d67c9-0e01-48ab-8091-2e0bf103655d" (UID: "be3d67c9-0e01-48ab-8091-2e0bf103655d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.731284 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-kube-api-access-zgt8s" (OuterVolumeSpecName: "kube-api-access-zgt8s") pod "a5f25640-5527-4caa-a290-cbb3bbfc5e0b" (UID: "a5f25640-5527-4caa-a290-cbb3bbfc5e0b"). InnerVolumeSpecName "kube-api-access-zgt8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.732143 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a5f25640-5527-4caa-a290-cbb3bbfc5e0b" (UID: "a5f25640-5527-4caa-a290-cbb3bbfc5e0b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827544 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827590 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/be3d67c9-0e01-48ab-8091-2e0bf103655d-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827603 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827614 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be3d67c9-0e01-48ab-8091-2e0bf103655d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827628 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827637 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827649 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgt8s\" (UniqueName: \"kubernetes.io/projected/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-kube-api-access-zgt8s\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827665 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a5f25640-5527-4caa-a290-cbb3bbfc5e0b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:47 crc kubenswrapper[4926]: I0312 18:07:47.827675 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw8k8\" (UniqueName: \"kubernetes.io/projected/be3d67c9-0e01-48ab-8091-2e0bf103655d-kube-api-access-mw8k8\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.016159 4926 generic.go:334] "Generic (PLEG): container finished" podID="a5f25640-5527-4caa-a290-cbb3bbfc5e0b" containerID="4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2" exitCode=0 Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.016269 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.016266 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" event={"ID":"a5f25640-5527-4caa-a290-cbb3bbfc5e0b","Type":"ContainerDied","Data":"4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2"} Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.016969 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74548584bf-x85ch" event={"ID":"a5f25640-5527-4caa-a290-cbb3bbfc5e0b","Type":"ContainerDied","Data":"9b5a165d0fdbcf22977eee75175ce438f8c21d2c260b57e0ebed1521f9e89fb5"} Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.017006 4926 scope.go:117] "RemoveContainer" containerID="4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.018921 4926 generic.go:334] "Generic (PLEG): container finished" podID="be3d67c9-0e01-48ab-8091-2e0bf103655d" containerID="19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128" exitCode=0 Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.018976 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" event={"ID":"be3d67c9-0e01-48ab-8091-2e0bf103655d","Type":"ContainerDied","Data":"19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128"} Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.019041 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" event={"ID":"be3d67c9-0e01-48ab-8091-2e0bf103655d","Type":"ContainerDied","Data":"581ff73fc13e04cb81a4a9c9a97934c14adf3bf26fd03837864c6f60ecb97184"} Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.019242 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.049462 4926 scope.go:117] "RemoveContainer" containerID="4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2" Mar 12 18:07:48 crc kubenswrapper[4926]: E0312 18:07:48.050127 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2\": container with ID starting with 4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2 not found: ID does not exist" containerID="4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.050166 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2"} err="failed to get container status \"4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2\": rpc error: code = NotFound desc = could not find container \"4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2\": container with ID starting with 4549267ab7af25c6d9ee4556186f3dd19ae56cb72fc8f13eb3af40054eada3f2 not found: ID does not exist" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.050190 4926 scope.go:117] "RemoveContainer" containerID="19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.062321 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74548584bf-x85ch"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.065419 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74548584bf-x85ch"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.073949 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.077591 4926 scope.go:117] "RemoveContainer" containerID="19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128" Mar 12 18:07:48 crc kubenswrapper[4926]: E0312 18:07:48.078276 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128\": container with ID starting with 19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128 not found: ID does not exist" containerID="19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.078334 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128"} err="failed to get container status \"19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128\": rpc error: code = NotFound desc = could not find container \"19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128\": container with ID starting with 19399e69849a8c08401404aad3ecf5af11661e7750f96f5570cd0700cde99128 not found: ID does not exist" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.079487 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b5767577f-5znp2"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.500239 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f25640-5527-4caa-a290-cbb3bbfc5e0b" path="/var/lib/kubelet/pods/a5f25640-5527-4caa-a290-cbb3bbfc5e0b/volumes" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.501090 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be3d67c9-0e01-48ab-8091-2e0bf103655d" path="/var/lib/kubelet/pods/be3d67c9-0e01-48ab-8091-2e0bf103655d/volumes" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.821599 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-cfc589789-g48s2"] Mar 12 18:07:48 crc kubenswrapper[4926]: E0312 18:07:48.821885 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f25640-5527-4caa-a290-cbb3bbfc5e0b" containerName="controller-manager" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.821899 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f25640-5527-4caa-a290-cbb3bbfc5e0b" containerName="controller-manager" Mar 12 18:07:48 crc kubenswrapper[4926]: E0312 18:07:48.821941 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be3d67c9-0e01-48ab-8091-2e0bf103655d" containerName="route-controller-manager" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.821950 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="be3d67c9-0e01-48ab-8091-2e0bf103655d" containerName="route-controller-manager" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.822054 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f25640-5527-4caa-a290-cbb3bbfc5e0b" containerName="controller-manager" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.822066 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="be3d67c9-0e01-48ab-8091-2e0bf103655d" containerName="route-controller-manager" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.822534 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.827188 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.827959 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.828618 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.828688 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.828855 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.829997 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.832881 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.834608 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.847200 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.847595 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cfc589789-g48s2"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.848214 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.848492 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9hgc\" (UniqueName: \"kubernetes.io/projected/439f4a23-9447-4999-b30e-7aa07f2b1553-kube-api-access-w9hgc\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.848754 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/439f4a23-9447-4999-b30e-7aa07f2b1553-serving-cert\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.849124 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbx6w\" (UniqueName: \"kubernetes.io/projected/0bcd8e23-295b-4e93-a842-3e25d10c10ef-kube-api-access-dbx6w\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.849368 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bcd8e23-295b-4e93-a842-3e25d10c10ef-client-ca\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.849612 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-config\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.849884 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bcd8e23-295b-4e93-a842-3e25d10c10ef-serving-cert\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.850067 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-client-ca\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.850414 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bcd8e23-295b-4e93-a842-3e25d10c10ef-config\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.849701 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.850606 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.850661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-proxy-ca-bundles\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.851597 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.853619 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.862374 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.862560 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.927072 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hmn6x"] Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.927402 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hmn6x" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="registry-server" containerID="cri-o://9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167" gracePeriod=2 Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.955674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbx6w\" (UniqueName: \"kubernetes.io/projected/0bcd8e23-295b-4e93-a842-3e25d10c10ef-kube-api-access-dbx6w\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.956046 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bcd8e23-295b-4e93-a842-3e25d10c10ef-client-ca\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.956182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-config\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.957032 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0bcd8e23-295b-4e93-a842-3e25d10c10ef-client-ca\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.957603 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bcd8e23-295b-4e93-a842-3e25d10c10ef-serving-cert\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.957655 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-client-ca\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.957679 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bcd8e23-295b-4e93-a842-3e25d10c10ef-config\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.958061 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-proxy-ca-bundles\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.958458 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-config\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.959745 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9hgc\" (UniqueName: \"kubernetes.io/projected/439f4a23-9447-4999-b30e-7aa07f2b1553-kube-api-access-w9hgc\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.959802 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/439f4a23-9447-4999-b30e-7aa07f2b1553-serving-cert\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.960247 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bcd8e23-295b-4e93-a842-3e25d10c10ef-config\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.960262 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-client-ca\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.960286 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/439f4a23-9447-4999-b30e-7aa07f2b1553-proxy-ca-bundles\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.965216 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0bcd8e23-295b-4e93-a842-3e25d10c10ef-serving-cert\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.965949 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/439f4a23-9447-4999-b30e-7aa07f2b1553-serving-cert\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.979591 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbx6w\" (UniqueName: \"kubernetes.io/projected/0bcd8e23-295b-4e93-a842-3e25d10c10ef-kube-api-access-dbx6w\") pod \"route-controller-manager-5499f549c8-qpmc8\" (UID: \"0bcd8e23-295b-4e93-a842-3e25d10c10ef\") " pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:48 crc kubenswrapper[4926]: I0312 18:07:48.988752 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9hgc\" (UniqueName: \"kubernetes.io/projected/439f4a23-9447-4999-b30e-7aa07f2b1553-kube-api-access-w9hgc\") pod \"controller-manager-cfc589789-g48s2\" (UID: \"439f4a23-9447-4999-b30e-7aa07f2b1553\") " pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.157115 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.166633 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.347311 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.368371 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-catalog-content\") pod \"1f425571-9ce5-4fdc-9631-7683efa292aa\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.368484 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-utilities\") pod \"1f425571-9ce5-4fdc-9631-7683efa292aa\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.368550 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4msz\" (UniqueName: \"kubernetes.io/projected/1f425571-9ce5-4fdc-9631-7683efa292aa-kube-api-access-v4msz\") pod \"1f425571-9ce5-4fdc-9631-7683efa292aa\" (UID: \"1f425571-9ce5-4fdc-9631-7683efa292aa\") " Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.369604 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-utilities" (OuterVolumeSpecName: "utilities") pod "1f425571-9ce5-4fdc-9631-7683efa292aa" (UID: "1f425571-9ce5-4fdc-9631-7683efa292aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.375658 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f425571-9ce5-4fdc-9631-7683efa292aa-kube-api-access-v4msz" (OuterVolumeSpecName: "kube-api-access-v4msz") pod "1f425571-9ce5-4fdc-9631-7683efa292aa" (UID: "1f425571-9ce5-4fdc-9631-7683efa292aa"). InnerVolumeSpecName "kube-api-access-v4msz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.406661 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8"] Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.470315 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.470345 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4msz\" (UniqueName: \"kubernetes.io/projected/1f425571-9ce5-4fdc-9631-7683efa292aa-kube-api-access-v4msz\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.503676 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f425571-9ce5-4fdc-9631-7683efa292aa" (UID: "1f425571-9ce5-4fdc-9631-7683efa292aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.571818 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f425571-9ce5-4fdc-9631-7683efa292aa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:49 crc kubenswrapper[4926]: I0312 18:07:49.660874 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cfc589789-g48s2"] Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.034859 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" event={"ID":"439f4a23-9447-4999-b30e-7aa07f2b1553","Type":"ContainerStarted","Data":"c6cbe8ef6255982af6256393e3246146b85748ebe1583b5f0e9e9a8d5abccf68"} Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.034926 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" event={"ID":"439f4a23-9447-4999-b30e-7aa07f2b1553","Type":"ContainerStarted","Data":"371829af9ec1fd277ceecbaffa5e067453dc8db27bb4cad207e9ce29d2560cda"} Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.035254 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.037149 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" event={"ID":"0bcd8e23-295b-4e93-a842-3e25d10c10ef","Type":"ContainerStarted","Data":"db0fe8825514367fe787d924830623f1474b74857c21979b15bf4cb6d0ccb0ed"} Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.037186 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" event={"ID":"0bcd8e23-295b-4e93-a842-3e25d10c10ef","Type":"ContainerStarted","Data":"46ea61a8c49a4b2f157f901e95f006ea70fa8ba384f61f1b6661bfc3e219c26c"} Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.037371 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.054689 4926 generic.go:334] "Generic (PLEG): container finished" podID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerID="9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167" exitCode=0 Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.054752 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmn6x" event={"ID":"1f425571-9ce5-4fdc-9631-7683efa292aa","Type":"ContainerDied","Data":"9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167"} Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.054787 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmn6x" event={"ID":"1f425571-9ce5-4fdc-9631-7683efa292aa","Type":"ContainerDied","Data":"d295389691829deab9da8e29d0c26dfd809e13bacff8864270d0d1a8c5b14dcd"} Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.054813 4926 scope.go:117] "RemoveContainer" containerID="9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.054815 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmn6x" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.068821 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.091674 4926 scope.go:117] "RemoveContainer" containerID="9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.115633 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-cfc589789-g48s2" podStartSLOduration=3.115610955 podStartE2EDuration="3.115610955s" podCreationTimestamp="2026-03-12 18:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:50.081279221 +0000 UTC m=+310.449905554" watchObservedRunningTime="2026-03-12 18:07:50.115610955 +0000 UTC m=+310.484237288" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.129023 4926 scope.go:117] "RemoveContainer" containerID="3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.153477 4926 scope.go:117] "RemoveContainer" containerID="9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167" Mar 12 18:07:50 crc kubenswrapper[4926]: E0312 18:07:50.155842 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167\": container with ID starting with 9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167 not found: ID does not exist" containerID="9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.155890 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167"} err="failed to get container status \"9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167\": rpc error: code = NotFound desc = could not find container \"9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167\": container with ID starting with 9953fa9c8ae4779c66f50a9da2e5055796da72263bfafc670cc521d8eb8da167 not found: ID does not exist" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.155915 4926 scope.go:117] "RemoveContainer" containerID="9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429" Mar 12 18:07:50 crc kubenswrapper[4926]: E0312 18:07:50.156226 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429\": container with ID starting with 9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429 not found: ID does not exist" containerID="9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.156313 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429"} err="failed to get container status \"9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429\": rpc error: code = NotFound desc = could not find container \"9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429\": container with ID starting with 9a932193e2b455fb9379d7e34c8f9171536a3e43595b8c2b4822103bfb25a429 not found: ID does not exist" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.156380 4926 scope.go:117] "RemoveContainer" containerID="3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a" Mar 12 18:07:50 crc kubenswrapper[4926]: E0312 18:07:50.159131 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a\": container with ID starting with 3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a not found: ID does not exist" containerID="3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.159306 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a"} err="failed to get container status \"3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a\": rpc error: code = NotFound desc = could not find container \"3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a\": container with ID starting with 3e95babad73f5786563a2d69b70a26fb937a9b6ccb630b521fa9357c578cba1a not found: ID does not exist" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.162760 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" podStartSLOduration=3.162747862 podStartE2EDuration="3.162747862s" podCreationTimestamp="2026-03-12 18:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:07:50.160668946 +0000 UTC m=+310.529295289" watchObservedRunningTime="2026-03-12 18:07:50.162747862 +0000 UTC m=+310.531374195" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.186934 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hmn6x"] Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.190660 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hmn6x"] Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.198199 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5499f549c8-qpmc8" Mar 12 18:07:50 crc kubenswrapper[4926]: I0312 18:07:50.495325 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" path="/var/lib/kubelet/pods/1f425571-9ce5-4fdc-9631-7683efa292aa/volumes" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.504717 4926 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.505203 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="extract-content" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.505218 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="extract-content" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.505231 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="registry-server" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.505239 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="registry-server" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.505257 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="extract-utilities" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.505266 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="extract-utilities" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.505404 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f425571-9ce5-4fdc-9631-7683efa292aa" containerName="registry-server" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.505827 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.506481 4926 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.506808 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b" gracePeriod=15 Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.506876 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808" gracePeriod=15 Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.506895 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319" gracePeriod=15 Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.506945 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4" gracePeriod=15 Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.506991 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882" gracePeriod=15 Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508643 4926 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508783 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508793 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508805 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508815 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508825 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508833 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508844 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508852 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508861 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508868 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508879 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508888 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508898 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508906 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508917 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508925 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.508936 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.508945 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509047 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509058 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509068 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509079 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509090 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509099 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509109 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509119 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.509222 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509231 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.509369 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.523563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.523682 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.523727 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.523968 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.524125 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.524222 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.524292 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.524369 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.559219 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.626944 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627514 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627589 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627618 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627638 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627664 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627681 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627707 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627782 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627827 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627851 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627873 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627905 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627927 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.627953 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.628022 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: I0312 18:07:53.855715 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:07:53 crc kubenswrapper[4926]: E0312 18:07:53.875952 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.238:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c2a4f6bb58173 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:07:53.874973043 +0000 UTC m=+314.243599386,LastTimestamp:2026-03-12 18:07:53.874973043 +0000 UTC m=+314.243599386,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.077624 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"66936a30ed9b3cad6731e6f16633aaba714831297c3f40433f5955a13bdac350"} Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.080215 4926 generic.go:334] "Generic (PLEG): container finished" podID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" containerID="988893c573b728d62d468c803dd6aa2333147e0f68cf77f807942af0ae0d7912" exitCode=0 Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.080267 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f","Type":"ContainerDied","Data":"988893c573b728d62d468c803dd6aa2333147e0f68cf77f807942af0ae0d7912"} Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.081254 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.081726 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.081875 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.082105 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.083373 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.084020 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319" exitCode=0 Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.084043 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882" exitCode=0 Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.084052 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808" exitCode=0 Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.084063 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4" exitCode=2 Mar 12 18:07:54 crc kubenswrapper[4926]: I0312 18:07:54.084104 4926 scope.go:117] "RemoveContainer" containerID="21b030cc5a2a69caa8b11ed4cd0b4872399a6d989caa30ccaae9b5dcf68e5eab" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.094109 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.097922 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d"} Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.098630 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.100144 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.448862 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.449790 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.450303 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.552956 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kubelet-dir\") pod \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.553055 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" (UID: "63c14bbd-6ba6-42c2-9e94-bf3a6f68500f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.553068 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kube-api-access\") pod \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.553148 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-var-lock\") pod \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\" (UID: \"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f\") " Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.553387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-var-lock" (OuterVolumeSpecName: "var-lock") pod "63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" (UID: "63c14bbd-6ba6-42c2-9e94-bf3a6f68500f"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.553522 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.561838 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" (UID: "63c14bbd-6ba6-42c2-9e94-bf3a6f68500f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.655599 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:55 crc kubenswrapper[4926]: I0312 18:07:55.656105 4926 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/63c14bbd-6ba6-42c2-9e94-bf3a6f68500f-var-lock\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.110853 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.114187 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b" exitCode=0 Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.117775 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.117849 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"63c14bbd-6ba6-42c2-9e94-bf3a6f68500f","Type":"ContainerDied","Data":"491a64d2c203ce09a50339566c321ae1dad16f74cdb93b33b26da5d1f82b4544"} Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.117922 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="491a64d2c203ce09a50339566c321ae1dad16f74cdb93b33b26da5d1f82b4544" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.147477 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.148406 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.424112 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.425211 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.425818 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.427374 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.427912 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468034 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468099 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468131 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468194 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468224 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468345 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468777 4926 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468821 4926 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.468842 4926 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:07:56 crc kubenswrapper[4926]: I0312 18:07:56.497762 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 12 18:07:56 crc kubenswrapper[4926]: E0312 18:07:56.541603 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.238:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c2a4f6bb58173 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:07:53.874973043 +0000 UTC m=+314.243599386,LastTimestamp:2026-03-12 18:07:53.874973043 +0000 UTC m=+314.243599386,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.126461 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.127370 4926 scope.go:117] "RemoveContainer" containerID="3d571ee532648a1c75519f9efc8effdd164f98979e9ab9d53610c940b0200319" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.127548 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.128142 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.128890 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.129193 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.131111 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.131418 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.131805 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.141974 4926 scope.go:117] "RemoveContainer" containerID="0905b9bfa33c6d9a362e6ef466636752b4bfa285f62a2376b909d31a96731882" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.158349 4926 scope.go:117] "RemoveContainer" containerID="077afd62dc8a90f869f162035ff5e84edf160a3105daf1a871634d10d13f9808" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.173281 4926 scope.go:117] "RemoveContainer" containerID="30d6c9fcb202b72e8db949c2ad4c55ba835f1e2b6f10db291360256e47884bf4" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.189337 4926 scope.go:117] "RemoveContainer" containerID="9c94763fccdf84de07158693ecdaf7781b6d2e8deae62cb9bd0d2bc8ffbddd7b" Mar 12 18:07:57 crc kubenswrapper[4926]: I0312 18:07:57.207414 4926 scope.go:117] "RemoveContainer" containerID="274f012f33be81fa551881a3695deba9e4195c945b1af81f04f11d56d1a3ea17" Mar 12 18:08:00 crc kubenswrapper[4926]: I0312 18:08:00.492585 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:00 crc kubenswrapper[4926]: I0312 18:08:00.494224 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.186856 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.187549 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.187875 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.188193 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.188668 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:02 crc kubenswrapper[4926]: I0312 18:08:02.188751 4926 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.189135 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="200ms" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.390193 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="400ms" Mar 12 18:08:02 crc kubenswrapper[4926]: E0312 18:08:02.791633 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="800ms" Mar 12 18:08:03 crc kubenswrapper[4926]: E0312 18:08:03.593565 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="1.6s" Mar 12 18:08:05 crc kubenswrapper[4926]: E0312 18:08:05.194508 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="3.2s" Mar 12 18:08:05 crc kubenswrapper[4926]: I0312 18:08:05.489059 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:05 crc kubenswrapper[4926]: I0312 18:08:05.490130 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:05 crc kubenswrapper[4926]: I0312 18:08:05.490595 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:05 crc kubenswrapper[4926]: I0312 18:08:05.507163 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:05 crc kubenswrapper[4926]: I0312 18:08:05.507678 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:05 crc kubenswrapper[4926]: E0312 18:08:05.508346 4926 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:05 crc kubenswrapper[4926]: I0312 18:08:05.508977 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:06 crc kubenswrapper[4926]: I0312 18:08:06.193067 4926 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="94e0bc894b14dc9ff48b9251eaa507b5e6745e7b7b2af0a3fca692cb764bfb7d" exitCode=0 Mar 12 18:08:06 crc kubenswrapper[4926]: I0312 18:08:06.193154 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"94e0bc894b14dc9ff48b9251eaa507b5e6745e7b7b2af0a3fca692cb764bfb7d"} Mar 12 18:08:06 crc kubenswrapper[4926]: I0312 18:08:06.193195 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2a62f4e45010887aed487bd146dda0ec82a48298f0bb54b58f1a02b9e6c8235b"} Mar 12 18:08:06 crc kubenswrapper[4926]: I0312 18:08:06.193631 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:06 crc kubenswrapper[4926]: I0312 18:08:06.193647 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:06 crc kubenswrapper[4926]: I0312 18:08:06.194433 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:06 crc kubenswrapper[4926]: E0312 18:08:06.194510 4926 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:06 crc kubenswrapper[4926]: I0312 18:08:06.194864 4926 status_manager.go:851] "Failed to get status for pod" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 12 18:08:06 crc kubenswrapper[4926]: E0312 18:08:06.543788 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.238:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c2a4f6bb58173 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 18:07:53.874973043 +0000 UTC m=+314.243599386,LastTimestamp:2026-03-12 18:07:53.874973043 +0000 UTC m=+314.243599386,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.218108 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.220676 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.220728 4926 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0476f8d782654dd0fc67232862117105145e03f9a9495fb7b7015c28dac5c4dc" exitCode=1 Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.220807 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0476f8d782654dd0fc67232862117105145e03f9a9495fb7b7015c28dac5c4dc"} Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.221202 4926 scope.go:117] "RemoveContainer" containerID="0476f8d782654dd0fc67232862117105145e03f9a9495fb7b7015c28dac5c4dc" Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.224561 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1075a64629b41faf6d84d7d274a4f2a3a4d4ea1b331e23003a01b3929fe7f326"} Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.224610 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"42d1a9d7ce46fada349e899645be2abf96f39c48643131be836bb1e6ba317eb8"} Mar 12 18:08:07 crc kubenswrapper[4926]: I0312 18:08:07.224625 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9c9f907cab99bf962950a3b0868cccac7d05aa470a5a7ee87e0adc217fbed488"} Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.233098 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.234861 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.234937 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c5f31c459da9f76e8cc69a726d7ff85128963c66fb1d5ef95af8a6621a8c3286"} Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.238087 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ceee911b542f90d3c7df26ffa0b59ee92dfa501fe1fc246b3066939c805fecfd"} Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.238419 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.238490 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.238784 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d3d56189bb24e1bd56ef327ad4e2750cc591d838202e17d442d7bf7872fb8746"} Mar 12 18:08:08 crc kubenswrapper[4926]: I0312 18:08:08.238822 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:09 crc kubenswrapper[4926]: I0312 18:08:09.324183 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:08:10 crc kubenswrapper[4926]: I0312 18:08:10.509330 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:10 crc kubenswrapper[4926]: I0312 18:08:10.509361 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:10 crc kubenswrapper[4926]: I0312 18:08:10.516764 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:12 crc kubenswrapper[4926]: I0312 18:08:12.438978 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:08:12 crc kubenswrapper[4926]: I0312 18:08:12.439512 4926 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 12 18:08:12 crc kubenswrapper[4926]: I0312 18:08:12.439608 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 12 18:08:13 crc kubenswrapper[4926]: I0312 18:08:13.270523 4926 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:14 crc kubenswrapper[4926]: I0312 18:08:14.276401 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:14 crc kubenswrapper[4926]: I0312 18:08:14.276717 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:14 crc kubenswrapper[4926]: I0312 18:08:14.280407 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:14 crc kubenswrapper[4926]: I0312 18:08:14.283190 4926 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="34d15346-5887-460e-8af4-28f9329010dd" Mar 12 18:08:15 crc kubenswrapper[4926]: I0312 18:08:15.282095 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:15 crc kubenswrapper[4926]: I0312 18:08:15.282158 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aeb621bb-05ee-456b-b869-1cdd14184ad1" Mar 12 18:08:20 crc kubenswrapper[4926]: I0312 18:08:20.520158 4926 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="34d15346-5887-460e-8af4-28f9329010dd" Mar 12 18:08:22 crc kubenswrapper[4926]: I0312 18:08:22.135629 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 12 18:08:22 crc kubenswrapper[4926]: I0312 18:08:22.445488 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:08:22 crc kubenswrapper[4926]: I0312 18:08:22.453311 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 18:08:22 crc kubenswrapper[4926]: I0312 18:08:22.960384 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 12 18:08:23 crc kubenswrapper[4926]: I0312 18:08:23.710113 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 12 18:08:23 crc kubenswrapper[4926]: I0312 18:08:23.753630 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.071087 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.349672 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.456714 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.638165 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.673287 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.718933 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.777401 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 12 18:08:24 crc kubenswrapper[4926]: I0312 18:08:24.937374 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 12 18:08:25 crc kubenswrapper[4926]: I0312 18:08:25.013003 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 18:08:25 crc kubenswrapper[4926]: I0312 18:08:25.209140 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 12 18:08:25 crc kubenswrapper[4926]: I0312 18:08:25.596951 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 12 18:08:25 crc kubenswrapper[4926]: I0312 18:08:25.862591 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 12 18:08:25 crc kubenswrapper[4926]: I0312 18:08:25.888256 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 12 18:08:25 crc kubenswrapper[4926]: I0312 18:08:25.975221 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.060854 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.179525 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.195996 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.212956 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.389690 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.401120 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.528720 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.761796 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 12 18:08:26 crc kubenswrapper[4926]: I0312 18:08:26.990941 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.001897 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.043847 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.060242 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.378493 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.497983 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.664406 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.709465 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.744197 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.782347 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.803861 4926 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.805068 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 18:08:27 crc kubenswrapper[4926]: I0312 18:08:27.843885 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.080787 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.146298 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.192257 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.310551 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.403796 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.435585 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.468757 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.485703 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.509427 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.642791 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.760230 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.761530 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.773770 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.776681 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.783352 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.802091 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.822343 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.854694 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.902565 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 12 18:08:28 crc kubenswrapper[4926]: I0312 18:08:28.935750 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.049220 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.067136 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.111327 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.122721 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.283290 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.283939 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.318817 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.335282 4926 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.341804 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.354677 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.475862 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.537687 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.593882 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.665265 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.674777 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.861869 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 12 18:08:29 crc kubenswrapper[4926]: I0312 18:08:29.942571 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.019410 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.029019 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.052454 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.059891 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.172241 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.184224 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.203389 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.252472 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.264205 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.347300 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.364580 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.388624 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.466600 4926 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.467869 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=37.46785532 podStartE2EDuration="37.46785532s" podCreationTimestamp="2026-03-12 18:07:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:08:12.933649342 +0000 UTC m=+333.302275685" watchObservedRunningTime="2026-03-12 18:08:30.46785532 +0000 UTC m=+350.836481653" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.470839 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.470885 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.475388 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.487252 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.48723427 podStartE2EDuration="17.48723427s" podCreationTimestamp="2026-03-12 18:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:08:30.48657967 +0000 UTC m=+350.855206013" watchObservedRunningTime="2026-03-12 18:08:30.48723427 +0000 UTC m=+350.855860603" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.526452 4926 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.537614 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.643997 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.664317 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.689498 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.762539 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.811532 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.851614 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.974360 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 12 18:08:30 crc kubenswrapper[4926]: I0312 18:08:30.977259 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.030194 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.034587 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.167709 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.174611 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.192185 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.250584 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.360673 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.422076 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.450019 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.455790 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.457032 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.479532 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.517215 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.517554 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.524283 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.539944 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.543162 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.550790 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.635554 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.799433 4926 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.799655 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.804160 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.828056 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.837494 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.893392 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 12 18:08:31 crc kubenswrapper[4926]: I0312 18:08:31.997375 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.146141 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.196667 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.244546 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.305517 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.328090 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.369760 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.381359 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.420240 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.427285 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.463304 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.481939 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.491728 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.513966 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 12 18:08:32 crc kubenswrapper[4926]: I0312 18:08:32.685760 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.002580 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.013742 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.055415 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.195353 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.281217 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.304348 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.372490 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.385887 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.436363 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.440631 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.511662 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.521293 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.618110 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.620694 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.703763 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.733523 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.753188 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 12 18:08:33 crc kubenswrapper[4926]: I0312 18:08:33.809154 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.044202 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.083838 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.092484 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.164484 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.281424 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.293076 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.296240 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.319055 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.319425 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.323379 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.329143 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.439902 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.485825 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.549936 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.621992 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.636307 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.670051 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.703149 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.805911 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.807129 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 12 18:08:34 crc kubenswrapper[4926]: I0312 18:08:34.954996 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.009589 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.081860 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.188185 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.324837 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.354960 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.390831 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.472127 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.517956 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.624854 4926 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.625171 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d" gracePeriod=5 Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.666834 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.762309 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.795678 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 12 18:08:35 crc kubenswrapper[4926]: I0312 18:08:35.981857 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.050596 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.097050 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.159855 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.197471 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.326517 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.345194 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.349687 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.369529 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.436768 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.552250 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.573200 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.659617 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.662739 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.741101 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.746178 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.789587 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.804152 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.826949 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.830731 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.859666 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.944894 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 12 18:08:36 crc kubenswrapper[4926]: I0312 18:08:36.987417 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.003136 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.083670 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.111822 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.157638 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.404856 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.446706 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.549702 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.578531 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.603929 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.629622 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.744634 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.880364 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.964026 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 12 18:08:37 crc kubenswrapper[4926]: I0312 18:08:37.988672 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 12 18:08:38 crc kubenswrapper[4926]: I0312 18:08:38.006176 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 12 18:08:38 crc kubenswrapper[4926]: I0312 18:08:38.016154 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 12 18:08:38 crc kubenswrapper[4926]: I0312 18:08:38.089564 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 12 18:08:38 crc kubenswrapper[4926]: I0312 18:08:38.677662 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 12 18:08:38 crc kubenswrapper[4926]: I0312 18:08:38.755559 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 12 18:08:38 crc kubenswrapper[4926]: I0312 18:08:38.886309 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.184782 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.280676 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.378101 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.454958 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.461786 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.863286 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.871193 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.895709 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.898582 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 12 18:08:39 crc kubenswrapper[4926]: I0312 18:08:39.956542 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.032308 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.175852 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.301498 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.447492 4926 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.558360 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.597974 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.686957 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.750323 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.750420 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.831775 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.831863 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.831901 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.831927 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.831907 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.831946 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.832042 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.832086 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.832358 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.832657 4926 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.832688 4926 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.832700 4926 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.832711 4926 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.844661 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.933599 4926 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.940263 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555648-j9n25"] Mar 12 18:08:40 crc kubenswrapper[4926]: E0312 18:08:40.940471 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.940486 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 18:08:40 crc kubenswrapper[4926]: E0312 18:08:40.940502 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" containerName="installer" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.940508 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" containerName="installer" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.940585 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c14bbd-6ba6-42c2-9e94-bf3a6f68500f" containerName="installer" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.940596 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.940930 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555648-j9n25" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.944983 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.945123 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.950854 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:08:40 crc kubenswrapper[4926]: I0312 18:08:40.953008 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555648-j9n25"] Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.034767 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wkg7\" (UniqueName: \"kubernetes.io/projected/8dbfae2a-83a6-4586-a33c-b9bcfd4df092-kube-api-access-2wkg7\") pod \"auto-csr-approver-29555648-j9n25\" (UID: \"8dbfae2a-83a6-4586-a33c-b9bcfd4df092\") " pod="openshift-infra/auto-csr-approver-29555648-j9n25" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.135502 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wkg7\" (UniqueName: \"kubernetes.io/projected/8dbfae2a-83a6-4586-a33c-b9bcfd4df092-kube-api-access-2wkg7\") pod \"auto-csr-approver-29555648-j9n25\" (UID: \"8dbfae2a-83a6-4586-a33c-b9bcfd4df092\") " pod="openshift-infra/auto-csr-approver-29555648-j9n25" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.155465 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wkg7\" (UniqueName: \"kubernetes.io/projected/8dbfae2a-83a6-4586-a33c-b9bcfd4df092-kube-api-access-2wkg7\") pod \"auto-csr-approver-29555648-j9n25\" (UID: \"8dbfae2a-83a6-4586-a33c-b9bcfd4df092\") " pod="openshift-infra/auto-csr-approver-29555648-j9n25" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.254246 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555648-j9n25" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.462081 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.462131 4926 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d" exitCode=137 Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.462171 4926 scope.go:117] "RemoveContainer" containerID="3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.462290 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.497069 4926 scope.go:117] "RemoveContainer" containerID="3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d" Mar 12 18:08:41 crc kubenswrapper[4926]: E0312 18:08:41.497944 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d\": container with ID starting with 3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d not found: ID does not exist" containerID="3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.497991 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d"} err="failed to get container status \"3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d\": rpc error: code = NotFound desc = could not find container \"3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d\": container with ID starting with 3fa55b458c434da936089801bf146dd071861b6c00213ac408d53f6b45376d4d not found: ID does not exist" Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.756781 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555648-j9n25"] Mar 12 18:08:41 crc kubenswrapper[4926]: I0312 18:08:41.833125 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.104895 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.468458 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555648-j9n25" event={"ID":"8dbfae2a-83a6-4586-a33c-b9bcfd4df092","Type":"ContainerStarted","Data":"4dd1bd1bd4a15c672854c7bab2a44c18a4e95ed9923be222520bdc3a11e519d2"} Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.496146 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.496458 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.506852 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.506890 4926 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="416312af-d40e-46e7-8345-9f214dc327c3" Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.510523 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.510559 4926 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="416312af-d40e-46e7-8345-9f214dc327c3" Mar 12 18:08:42 crc kubenswrapper[4926]: I0312 18:08:42.528774 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 12 18:08:43 crc kubenswrapper[4926]: I0312 18:08:43.476060 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555648-j9n25" event={"ID":"8dbfae2a-83a6-4586-a33c-b9bcfd4df092","Type":"ContainerStarted","Data":"57bd95669fead78f455d353e126bbd0157addbc0b0171c28ad86c1bdee789263"} Mar 12 18:08:43 crc kubenswrapper[4926]: I0312 18:08:43.486691 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555648-j9n25" podStartSLOduration=2.070084461 podStartE2EDuration="3.486673164s" podCreationTimestamp="2026-03-12 18:08:40 +0000 UTC" firstStartedPulling="2026-03-12 18:08:41.753704053 +0000 UTC m=+362.122330386" lastFinishedPulling="2026-03-12 18:08:43.170292756 +0000 UTC m=+363.538919089" observedRunningTime="2026-03-12 18:08:43.486518339 +0000 UTC m=+363.855144672" watchObservedRunningTime="2026-03-12 18:08:43.486673164 +0000 UTC m=+363.855299497" Mar 12 18:08:44 crc kubenswrapper[4926]: I0312 18:08:44.481580 4926 generic.go:334] "Generic (PLEG): container finished" podID="8dbfae2a-83a6-4586-a33c-b9bcfd4df092" containerID="57bd95669fead78f455d353e126bbd0157addbc0b0171c28ad86c1bdee789263" exitCode=0 Mar 12 18:08:44 crc kubenswrapper[4926]: I0312 18:08:44.481625 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555648-j9n25" event={"ID":"8dbfae2a-83a6-4586-a33c-b9bcfd4df092","Type":"ContainerDied","Data":"57bd95669fead78f455d353e126bbd0157addbc0b0171c28ad86c1bdee789263"} Mar 12 18:08:45 crc kubenswrapper[4926]: I0312 18:08:45.836236 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555648-j9n25" Mar 12 18:08:45 crc kubenswrapper[4926]: I0312 18:08:45.893563 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wkg7\" (UniqueName: \"kubernetes.io/projected/8dbfae2a-83a6-4586-a33c-b9bcfd4df092-kube-api-access-2wkg7\") pod \"8dbfae2a-83a6-4586-a33c-b9bcfd4df092\" (UID: \"8dbfae2a-83a6-4586-a33c-b9bcfd4df092\") " Mar 12 18:08:45 crc kubenswrapper[4926]: I0312 18:08:45.900957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbfae2a-83a6-4586-a33c-b9bcfd4df092-kube-api-access-2wkg7" (OuterVolumeSpecName: "kube-api-access-2wkg7") pod "8dbfae2a-83a6-4586-a33c-b9bcfd4df092" (UID: "8dbfae2a-83a6-4586-a33c-b9bcfd4df092"). InnerVolumeSpecName "kube-api-access-2wkg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:08:45 crc kubenswrapper[4926]: I0312 18:08:45.994330 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wkg7\" (UniqueName: \"kubernetes.io/projected/8dbfae2a-83a6-4586-a33c-b9bcfd4df092-kube-api-access-2wkg7\") on node \"crc\" DevicePath \"\"" Mar 12 18:08:46 crc kubenswrapper[4926]: I0312 18:08:46.498020 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555648-j9n25" event={"ID":"8dbfae2a-83a6-4586-a33c-b9bcfd4df092","Type":"ContainerDied","Data":"4dd1bd1bd4a15c672854c7bab2a44c18a4e95ed9923be222520bdc3a11e519d2"} Mar 12 18:08:46 crc kubenswrapper[4926]: I0312 18:08:46.498105 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dd1bd1bd4a15c672854c7bab2a44c18a4e95ed9923be222520bdc3a11e519d2" Mar 12 18:08:46 crc kubenswrapper[4926]: I0312 18:08:46.498209 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555648-j9n25" Mar 12 18:09:00 crc kubenswrapper[4926]: I0312 18:09:00.590978 4926 generic.go:334] "Generic (PLEG): container finished" podID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerID="c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec" exitCode=0 Mar 12 18:09:00 crc kubenswrapper[4926]: I0312 18:09:00.591172 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" event={"ID":"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b","Type":"ContainerDied","Data":"c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec"} Mar 12 18:09:00 crc kubenswrapper[4926]: I0312 18:09:00.592628 4926 scope.go:117] "RemoveContainer" containerID="c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec" Mar 12 18:09:01 crc kubenswrapper[4926]: I0312 18:09:01.602865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" event={"ID":"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b","Type":"ContainerStarted","Data":"356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91"} Mar 12 18:09:01 crc kubenswrapper[4926]: I0312 18:09:01.603294 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:09:01 crc kubenswrapper[4926]: I0312 18:09:01.605690 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:09:56 crc kubenswrapper[4926]: I0312 18:09:56.817733 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:09:56 crc kubenswrapper[4926]: I0312 18:09:56.818740 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.780324 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4xt9k"] Mar 12 18:09:59 crc kubenswrapper[4926]: E0312 18:09:59.780651 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbfae2a-83a6-4586-a33c-b9bcfd4df092" containerName="oc" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.780670 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbfae2a-83a6-4586-a33c-b9bcfd4df092" containerName="oc" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.780839 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbfae2a-83a6-4586-a33c-b9bcfd4df092" containerName="oc" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.781357 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.804242 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4xt9k"] Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962146 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61de267b-cc52-4384-a66f-e8c2f5ddac13-trusted-ca\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962205 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61de267b-cc52-4384-a66f-e8c2f5ddac13-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962323 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61de267b-cc52-4384-a66f-e8c2f5ddac13-registry-certificates\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962365 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-registry-tls\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962404 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx9dg\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-kube-api-access-bx9dg\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962455 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-bound-sa-token\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962531 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.962619 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61de267b-cc52-4384-a66f-e8c2f5ddac13-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:09:59 crc kubenswrapper[4926]: I0312 18:09:59.986320 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.063412 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-registry-tls\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.063697 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61de267b-cc52-4384-a66f-e8c2f5ddac13-registry-certificates\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.063727 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx9dg\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-kube-api-access-bx9dg\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.063747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-bound-sa-token\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.063788 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61de267b-cc52-4384-a66f-e8c2f5ddac13-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.063822 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61de267b-cc52-4384-a66f-e8c2f5ddac13-trusted-ca\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.063852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61de267b-cc52-4384-a66f-e8c2f5ddac13-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.064432 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/61de267b-cc52-4384-a66f-e8c2f5ddac13-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.064977 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/61de267b-cc52-4384-a66f-e8c2f5ddac13-trusted-ca\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.065017 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/61de267b-cc52-4384-a66f-e8c2f5ddac13-registry-certificates\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.068745 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-registry-tls\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.069610 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/61de267b-cc52-4384-a66f-e8c2f5ddac13-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.088688 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-bound-sa-token\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.088727 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx9dg\" (UniqueName: \"kubernetes.io/projected/61de267b-cc52-4384-a66f-e8c2f5ddac13-kube-api-access-bx9dg\") pod \"image-registry-66df7c8f76-4xt9k\" (UID: \"61de267b-cc52-4384-a66f-e8c2f5ddac13\") " pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.104690 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.127811 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555650-t22sq"] Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.128572 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555650-t22sq" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.130667 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.130922 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.131015 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.142026 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555650-t22sq"] Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.266048 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgl86\" (UniqueName: \"kubernetes.io/projected/5a5ff64f-2478-4592-8a08-fb47a40a8de5-kube-api-access-cgl86\") pod \"auto-csr-approver-29555650-t22sq\" (UID: \"5a5ff64f-2478-4592-8a08-fb47a40a8de5\") " pod="openshift-infra/auto-csr-approver-29555650-t22sq" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.286381 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4xt9k"] Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.367219 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgl86\" (UniqueName: \"kubernetes.io/projected/5a5ff64f-2478-4592-8a08-fb47a40a8de5-kube-api-access-cgl86\") pod \"auto-csr-approver-29555650-t22sq\" (UID: \"5a5ff64f-2478-4592-8a08-fb47a40a8de5\") " pod="openshift-infra/auto-csr-approver-29555650-t22sq" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.389122 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgl86\" (UniqueName: \"kubernetes.io/projected/5a5ff64f-2478-4592-8a08-fb47a40a8de5-kube-api-access-cgl86\") pod \"auto-csr-approver-29555650-t22sq\" (UID: \"5a5ff64f-2478-4592-8a08-fb47a40a8de5\") " pod="openshift-infra/auto-csr-approver-29555650-t22sq" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.465053 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555650-t22sq" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.626055 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555650-t22sq"] Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.979589 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" event={"ID":"61de267b-cc52-4384-a66f-e8c2f5ddac13","Type":"ContainerStarted","Data":"a3ae85736cc481a99989d3e010a953e8a8bbf4438c14860f036e5e3ad93aaaa7"} Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.979976 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" event={"ID":"61de267b-cc52-4384-a66f-e8c2f5ddac13","Type":"ContainerStarted","Data":"3d9227b85d6bab68548e3175d5cc29dfcb8aacc3002008bb43918526dfec9c68"} Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.980006 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.980920 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555650-t22sq" event={"ID":"5a5ff64f-2478-4592-8a08-fb47a40a8de5","Type":"ContainerStarted","Data":"a7f386311505d74c38a0b78f7af734c7c38cca55a7266dfa47d2726163b7e6ff"} Mar 12 18:10:00 crc kubenswrapper[4926]: I0312 18:10:00.997289 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" podStartSLOduration=1.997270171 podStartE2EDuration="1.997270171s" podCreationTimestamp="2026-03-12 18:09:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:10:00.99722775 +0000 UTC m=+441.365854113" watchObservedRunningTime="2026-03-12 18:10:00.997270171 +0000 UTC m=+441.365896504" Mar 12 18:10:03 crc kubenswrapper[4926]: I0312 18:10:03.000049 4926 generic.go:334] "Generic (PLEG): container finished" podID="5a5ff64f-2478-4592-8a08-fb47a40a8de5" containerID="1d8805e76417daef23d168ee9dafcbdf012b728e6e9278f288005edef5f23d20" exitCode=0 Mar 12 18:10:03 crc kubenswrapper[4926]: I0312 18:10:03.000523 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555650-t22sq" event={"ID":"5a5ff64f-2478-4592-8a08-fb47a40a8de5","Type":"ContainerDied","Data":"1d8805e76417daef23d168ee9dafcbdf012b728e6e9278f288005edef5f23d20"} Mar 12 18:10:04 crc kubenswrapper[4926]: I0312 18:10:04.365604 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555650-t22sq" Mar 12 18:10:04 crc kubenswrapper[4926]: I0312 18:10:04.532571 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgl86\" (UniqueName: \"kubernetes.io/projected/5a5ff64f-2478-4592-8a08-fb47a40a8de5-kube-api-access-cgl86\") pod \"5a5ff64f-2478-4592-8a08-fb47a40a8de5\" (UID: \"5a5ff64f-2478-4592-8a08-fb47a40a8de5\") " Mar 12 18:10:04 crc kubenswrapper[4926]: I0312 18:10:04.538978 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a5ff64f-2478-4592-8a08-fb47a40a8de5-kube-api-access-cgl86" (OuterVolumeSpecName: "kube-api-access-cgl86") pod "5a5ff64f-2478-4592-8a08-fb47a40a8de5" (UID: "5a5ff64f-2478-4592-8a08-fb47a40a8de5"). InnerVolumeSpecName "kube-api-access-cgl86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:04 crc kubenswrapper[4926]: I0312 18:10:04.633722 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgl86\" (UniqueName: \"kubernetes.io/projected/5a5ff64f-2478-4592-8a08-fb47a40a8de5-kube-api-access-cgl86\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:05 crc kubenswrapper[4926]: I0312 18:10:05.020650 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555650-t22sq" event={"ID":"5a5ff64f-2478-4592-8a08-fb47a40a8de5","Type":"ContainerDied","Data":"a7f386311505d74c38a0b78f7af734c7c38cca55a7266dfa47d2726163b7e6ff"} Mar 12 18:10:05 crc kubenswrapper[4926]: I0312 18:10:05.021110 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7f386311505d74c38a0b78f7af734c7c38cca55a7266dfa47d2726163b7e6ff" Mar 12 18:10:05 crc kubenswrapper[4926]: I0312 18:10:05.020735 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555650-t22sq" Mar 12 18:10:20 crc kubenswrapper[4926]: I0312 18:10:20.109162 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4xt9k" Mar 12 18:10:20 crc kubenswrapper[4926]: I0312 18:10:20.199855 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g6fzt"] Mar 12 18:10:26 crc kubenswrapper[4926]: I0312 18:10:26.817641 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:10:26 crc kubenswrapper[4926]: I0312 18:10:26.818306 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.660055 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-565fl"] Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.660653 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-565fl" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="registry-server" containerID="cri-o://a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8" gracePeriod=30 Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.672547 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s4zf5"] Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.674158 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s4zf5" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="registry-server" containerID="cri-o://1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e" gracePeriod=30 Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.684001 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c68kr"] Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.684275 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" containerID="cri-o://356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91" gracePeriod=30 Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.692342 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz6qt"] Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.692568 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wz6qt" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="registry-server" containerID="cri-o://9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83" gracePeriod=30 Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.698882 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9gx5"] Mar 12 18:10:28 crc kubenswrapper[4926]: E0312 18:10:28.699179 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a5ff64f-2478-4592-8a08-fb47a40a8de5" containerName="oc" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.699199 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a5ff64f-2478-4592-8a08-fb47a40a8de5" containerName="oc" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.699334 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a5ff64f-2478-4592-8a08-fb47a40a8de5" containerName="oc" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.699826 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.702097 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zx9h7"] Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.702304 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zx9h7" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="registry-server" containerID="cri-o://2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5" gracePeriod=30 Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.711748 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/daeebaaf-6a69-436e-b341-36fae756599e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.711810 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/daeebaaf-6a69-436e-b341-36fae756599e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.711882 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmzgn\" (UniqueName: \"kubernetes.io/projected/daeebaaf-6a69-436e-b341-36fae756599e-kube-api-access-tmzgn\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.714574 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9gx5"] Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.812595 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/daeebaaf-6a69-436e-b341-36fae756599e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.813038 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/daeebaaf-6a69-436e-b341-36fae756599e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.813186 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmzgn\" (UniqueName: \"kubernetes.io/projected/daeebaaf-6a69-436e-b341-36fae756599e-kube-api-access-tmzgn\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.814511 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/daeebaaf-6a69-436e-b341-36fae756599e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.821805 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/daeebaaf-6a69-436e-b341-36fae756599e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:28 crc kubenswrapper[4926]: I0312 18:10:28.839533 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmzgn\" (UniqueName: \"kubernetes.io/projected/daeebaaf-6a69-436e-b341-36fae756599e-kube-api-access-tmzgn\") pod \"marketplace-operator-79b997595-d9gx5\" (UID: \"daeebaaf-6a69-436e-b341-36fae756599e\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.031546 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.078288 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.116170 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.148929 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.153341 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.154641 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.192647 4926 generic.go:334] "Generic (PLEG): container finished" podID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerID="9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83" exitCode=0 Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.192985 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz6qt" event={"ID":"702daa2d-851e-4c3d-be86-4f337b4462f7","Type":"ContainerDied","Data":"9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.193009 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wz6qt" event={"ID":"702daa2d-851e-4c3d-be86-4f337b4462f7","Type":"ContainerDied","Data":"1ca7001bbffd391adca6fddca7abaa8f8e5f4a6dec4a426455ef61c2d31fac2d"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.193026 4926 scope.go:117] "RemoveContainer" containerID="9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.193126 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wz6qt" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.198948 4926 generic.go:334] "Generic (PLEG): container finished" podID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerID="356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91" exitCode=0 Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.199019 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" event={"ID":"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b","Type":"ContainerDied","Data":"356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.199047 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" event={"ID":"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b","Type":"ContainerDied","Data":"7a6f3ba59fc7e8583a370a6e58f7ad7026a2677d263e98044721d21503f15589"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.199096 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-c68kr" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.208218 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerID="a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8" exitCode=0 Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.208277 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-565fl" event={"ID":"b5fe4032-6a1e-4c27-9471-fa53e044826e","Type":"ContainerDied","Data":"a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.208301 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-565fl" event={"ID":"b5fe4032-6a1e-4c27-9471-fa53e044826e","Type":"ContainerDied","Data":"2ea4f4bc9037632e2b998da8312ddbb98d227eabc09259bd968b95eab6d3b562"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.208350 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-565fl" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.212081 4926 generic.go:334] "Generic (PLEG): container finished" podID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerID="1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e" exitCode=0 Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.212134 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4zf5" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.212133 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4zf5" event={"ID":"b2a609cd-c298-4356-9ddf-a7f125b52938","Type":"ContainerDied","Data":"1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.212177 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4zf5" event={"ID":"b2a609cd-c298-4356-9ddf-a7f125b52938","Type":"ContainerDied","Data":"6da98836531143c0f0c9e5b0f54f53b5a8c52f32a98b1d55d2cd5e57c0ec7c9e"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.216097 4926 generic.go:334] "Generic (PLEG): container finished" podID="637236a6-6287-401d-a2cd-78713aa03176" containerID="2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5" exitCode=0 Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.216144 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx9h7" event={"ID":"637236a6-6287-401d-a2cd-78713aa03176","Type":"ContainerDied","Data":"2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.216196 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zx9h7" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.216207 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zx9h7" event={"ID":"637236a6-6287-401d-a2cd-78713aa03176","Type":"ContainerDied","Data":"b41dbdfd44a141e983428f384fe92d330459deb4bd26d614487e897aef5edb00"} Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.218081 4926 scope.go:117] "RemoveContainer" containerID="946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.226094 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-catalog-content\") pod \"b5fe4032-6a1e-4c27-9471-fa53e044826e\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.226144 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-catalog-content\") pod \"b2a609cd-c298-4356-9ddf-a7f125b52938\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.226179 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf7tf\" (UniqueName: \"kubernetes.io/projected/b2a609cd-c298-4356-9ddf-a7f125b52938-kube-api-access-wf7tf\") pod \"b2a609cd-c298-4356-9ddf-a7f125b52938\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.226205 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-utilities\") pod \"b2a609cd-c298-4356-9ddf-a7f125b52938\" (UID: \"b2a609cd-c298-4356-9ddf-a7f125b52938\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.226279 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-utilities\") pod \"b5fe4032-6a1e-4c27-9471-fa53e044826e\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.226302 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wxfn\" (UniqueName: \"kubernetes.io/projected/b5fe4032-6a1e-4c27-9471-fa53e044826e-kube-api-access-2wxfn\") pod \"b5fe4032-6a1e-4c27-9471-fa53e044826e\" (UID: \"b5fe4032-6a1e-4c27-9471-fa53e044826e\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.229633 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-utilities" (OuterVolumeSpecName: "utilities") pod "b2a609cd-c298-4356-9ddf-a7f125b52938" (UID: "b2a609cd-c298-4356-9ddf-a7f125b52938"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.229635 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-utilities" (OuterVolumeSpecName: "utilities") pod "b5fe4032-6a1e-4c27-9471-fa53e044826e" (UID: "b5fe4032-6a1e-4c27-9471-fa53e044826e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.234422 4926 scope.go:117] "RemoveContainer" containerID="7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.238491 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a609cd-c298-4356-9ddf-a7f125b52938-kube-api-access-wf7tf" (OuterVolumeSpecName: "kube-api-access-wf7tf") pod "b2a609cd-c298-4356-9ddf-a7f125b52938" (UID: "b2a609cd-c298-4356-9ddf-a7f125b52938"). InnerVolumeSpecName "kube-api-access-wf7tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.244632 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5fe4032-6a1e-4c27-9471-fa53e044826e-kube-api-access-2wxfn" (OuterVolumeSpecName: "kube-api-access-2wxfn") pod "b5fe4032-6a1e-4c27-9471-fa53e044826e" (UID: "b5fe4032-6a1e-4c27-9471-fa53e044826e"). InnerVolumeSpecName "kube-api-access-2wxfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.253645 4926 scope.go:117] "RemoveContainer" containerID="9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.254671 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83\": container with ID starting with 9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83 not found: ID does not exist" containerID="9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.254720 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83"} err="failed to get container status \"9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83\": rpc error: code = NotFound desc = could not find container \"9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83\": container with ID starting with 9865a7f38f113c6dde3c65a4e033ac09e71fbbe0591d4c816afba1cdad51af83 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.254778 4926 scope.go:117] "RemoveContainer" containerID="946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.256667 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b\": container with ID starting with 946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b not found: ID does not exist" containerID="946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.256708 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b"} err="failed to get container status \"946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b\": rpc error: code = NotFound desc = could not find container \"946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b\": container with ID starting with 946cba603efe15d0422595010c1330546d61cfb532c27ac70fff29c0e4b7670b not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.256733 4926 scope.go:117] "RemoveContainer" containerID="7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.257014 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa\": container with ID starting with 7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa not found: ID does not exist" containerID="7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.257047 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa"} err="failed to get container status \"7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa\": rpc error: code = NotFound desc = could not find container \"7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa\": container with ID starting with 7dc08056a58cbc293791a80a85351d04a9b2abbd923938014309511e4c4926aa not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.257067 4926 scope.go:117] "RemoveContainer" containerID="356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.268043 4926 scope.go:117] "RemoveContainer" containerID="c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.281475 4926 scope.go:117] "RemoveContainer" containerID="356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.281834 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91\": container with ID starting with 356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91 not found: ID does not exist" containerID="356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.281860 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91"} err="failed to get container status \"356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91\": rpc error: code = NotFound desc = could not find container \"356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91\": container with ID starting with 356f40aae500e05628cc7650828bff417d73534ec79c5865296b52de2f45db91 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.281877 4926 scope.go:117] "RemoveContainer" containerID="c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.282049 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec\": container with ID starting with c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec not found: ID does not exist" containerID="c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.282065 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec"} err="failed to get container status \"c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec\": rpc error: code = NotFound desc = could not find container \"c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec\": container with ID starting with c343107391a1e08d133d97a7a02674714d51675df1d543bc4b787c1e00fb0bec not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.282078 4926 scope.go:117] "RemoveContainer" containerID="a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.288121 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2a609cd-c298-4356-9ddf-a7f125b52938" (UID: "b2a609cd-c298-4356-9ddf-a7f125b52938"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.289166 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5fe4032-6a1e-4c27-9471-fa53e044826e" (UID: "b5fe4032-6a1e-4c27-9471-fa53e044826e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.301283 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9gx5"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.303493 4926 scope.go:117] "RemoveContainer" containerID="d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5" Mar 12 18:10:29 crc kubenswrapper[4926]: W0312 18:10:29.316857 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaeebaaf_6a69_436e_b341_36fae756599e.slice/crio-cb5bce34c6486bdd5b618e103d0c85368e73bb247bb05dc441733b928f66d9e4 WatchSource:0}: Error finding container cb5bce34c6486bdd5b618e103d0c85368e73bb247bb05dc441733b928f66d9e4: Status 404 returned error can't find the container with id cb5bce34c6486bdd5b618e103d0c85368e73bb247bb05dc441733b928f66d9e4 Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.320635 4926 scope.go:117] "RemoveContainer" containerID="fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327243 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdq7x\" (UniqueName: \"kubernetes.io/projected/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-kube-api-access-vdq7x\") pod \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327304 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-utilities\") pod \"702daa2d-851e-4c3d-be86-4f337b4462f7\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327330 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-catalog-content\") pod \"702daa2d-851e-4c3d-be86-4f337b4462f7\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327390 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-trusted-ca\") pod \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327409 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57ncv\" (UniqueName: \"kubernetes.io/projected/637236a6-6287-401d-a2cd-78713aa03176-kube-api-access-57ncv\") pod \"637236a6-6287-401d-a2cd-78713aa03176\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327495 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-utilities\") pod \"637236a6-6287-401d-a2cd-78713aa03176\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327513 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-operator-metrics\") pod \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\" (UID: \"16fff1b6-a4ca-4ea9-aaba-40e9c136f62b\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327570 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-catalog-content\") pod \"637236a6-6287-401d-a2cd-78713aa03176\" (UID: \"637236a6-6287-401d-a2cd-78713aa03176\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.327625 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwlr2\" (UniqueName: \"kubernetes.io/projected/702daa2d-851e-4c3d-be86-4f337b4462f7-kube-api-access-nwlr2\") pod \"702daa2d-851e-4c3d-be86-4f337b4462f7\" (UID: \"702daa2d-851e-4c3d-be86-4f337b4462f7\") " Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" (UID: "16fff1b6-a4ca-4ea9-aaba-40e9c136f62b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328728 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328767 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328777 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2a609cd-c298-4356-9ddf-a7f125b52938-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328786 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf7tf\" (UniqueName: \"kubernetes.io/projected/b2a609cd-c298-4356-9ddf-a7f125b52938-kube-api-access-wf7tf\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328796 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5fe4032-6a1e-4c27-9471-fa53e044826e-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328805 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wxfn\" (UniqueName: \"kubernetes.io/projected/b5fe4032-6a1e-4c27-9471-fa53e044826e-kube-api-access-2wxfn\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.328835 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.329303 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-utilities" (OuterVolumeSpecName: "utilities") pod "637236a6-6287-401d-a2cd-78713aa03176" (UID: "637236a6-6287-401d-a2cd-78713aa03176"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.330077 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" (UID: "16fff1b6-a4ca-4ea9-aaba-40e9c136f62b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.330527 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-kube-api-access-vdq7x" (OuterVolumeSpecName: "kube-api-access-vdq7x") pod "16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" (UID: "16fff1b6-a4ca-4ea9-aaba-40e9c136f62b"). InnerVolumeSpecName "kube-api-access-vdq7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.331388 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702daa2d-851e-4c3d-be86-4f337b4462f7-kube-api-access-nwlr2" (OuterVolumeSpecName: "kube-api-access-nwlr2") pod "702daa2d-851e-4c3d-be86-4f337b4462f7" (UID: "702daa2d-851e-4c3d-be86-4f337b4462f7"). InnerVolumeSpecName "kube-api-access-nwlr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.332267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-utilities" (OuterVolumeSpecName: "utilities") pod "702daa2d-851e-4c3d-be86-4f337b4462f7" (UID: "702daa2d-851e-4c3d-be86-4f337b4462f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.340505 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/637236a6-6287-401d-a2cd-78713aa03176-kube-api-access-57ncv" (OuterVolumeSpecName: "kube-api-access-57ncv") pod "637236a6-6287-401d-a2cd-78713aa03176" (UID: "637236a6-6287-401d-a2cd-78713aa03176"). InnerVolumeSpecName "kube-api-access-57ncv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.355272 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "702daa2d-851e-4c3d-be86-4f337b4462f7" (UID: "702daa2d-851e-4c3d-be86-4f337b4462f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.357411 4926 scope.go:117] "RemoveContainer" containerID="a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.357828 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8\": container with ID starting with a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8 not found: ID does not exist" containerID="a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.357870 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8"} err="failed to get container status \"a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8\": rpc error: code = NotFound desc = could not find container \"a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8\": container with ID starting with a7cffeb42c14a337a70a12925c231b7706bf9cf89e483b52294bc445995635e8 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.357900 4926 scope.go:117] "RemoveContainer" containerID="d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.358245 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5\": container with ID starting with d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5 not found: ID does not exist" containerID="d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.358291 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5"} err="failed to get container status \"d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5\": rpc error: code = NotFound desc = could not find container \"d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5\": container with ID starting with d0ec30d7572ff20ab50a53e18ca39227cc44a4942924ccee935f4244f66215a5 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.358304 4926 scope.go:117] "RemoveContainer" containerID="fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.358735 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8\": container with ID starting with fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8 not found: ID does not exist" containerID="fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.358763 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8"} err="failed to get container status \"fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8\": rpc error: code = NotFound desc = could not find container \"fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8\": container with ID starting with fb58a38d637092f295d67153f35c6b3f40a27bae368a932eb26c3f8584a612c8 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.358781 4926 scope.go:117] "RemoveContainer" containerID="1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.375416 4926 scope.go:117] "RemoveContainer" containerID="f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.389905 4926 scope.go:117] "RemoveContainer" containerID="12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.402607 4926 scope.go:117] "RemoveContainer" containerID="1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.403245 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e\": container with ID starting with 1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e not found: ID does not exist" containerID="1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.403339 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e"} err="failed to get container status \"1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e\": rpc error: code = NotFound desc = could not find container \"1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e\": container with ID starting with 1e7c10918fb2662f0c3c2d62f7e215a766fbc3071c1ee00eafa305241344c76e not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.403421 4926 scope.go:117] "RemoveContainer" containerID="f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.403808 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc\": container with ID starting with f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc not found: ID does not exist" containerID="f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.403916 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc"} err="failed to get container status \"f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc\": rpc error: code = NotFound desc = could not find container \"f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc\": container with ID starting with f38216f8913635b4032c7273cdec95f688341e0baf732831b6a09cf20c9441dc not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.403981 4926 scope.go:117] "RemoveContainer" containerID="12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.404253 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447\": container with ID starting with 12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447 not found: ID does not exist" containerID="12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.404327 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447"} err="failed to get container status \"12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447\": rpc error: code = NotFound desc = could not find container \"12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447\": container with ID starting with 12072b6b2001c513ccb9e0bfcc29df831e119de1d1ab215530bbc5c9682cd447 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.404400 4926 scope.go:117] "RemoveContainer" containerID="2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.420041 4926 scope.go:117] "RemoveContainer" containerID="599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.430120 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.430153 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.430164 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwlr2\" (UniqueName: \"kubernetes.io/projected/702daa2d-851e-4c3d-be86-4f337b4462f7-kube-api-access-nwlr2\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.430174 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdq7x\" (UniqueName: \"kubernetes.io/projected/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b-kube-api-access-vdq7x\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.430183 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.430192 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/702daa2d-851e-4c3d-be86-4f337b4462f7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.430200 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57ncv\" (UniqueName: \"kubernetes.io/projected/637236a6-6287-401d-a2cd-78713aa03176-kube-api-access-57ncv\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.443705 4926 scope.go:117] "RemoveContainer" containerID="76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.460133 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "637236a6-6287-401d-a2cd-78713aa03176" (UID: "637236a6-6287-401d-a2cd-78713aa03176"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.468230 4926 scope.go:117] "RemoveContainer" containerID="2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.469625 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5\": container with ID starting with 2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5 not found: ID does not exist" containerID="2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.469663 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5"} err="failed to get container status \"2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5\": rpc error: code = NotFound desc = could not find container \"2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5\": container with ID starting with 2c8c0fa29e4cec7c2b5e010a9e7a63d04b2f8d113b9bc5720d7a9f2830e056e5 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.469688 4926 scope.go:117] "RemoveContainer" containerID="599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.470304 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb\": container with ID starting with 599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb not found: ID does not exist" containerID="599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.470332 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb"} err="failed to get container status \"599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb\": rpc error: code = NotFound desc = could not find container \"599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb\": container with ID starting with 599c070b0f1e1e00eef408d7433fad843523b69117524e7dec85b71cf79b99cb not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.470352 4926 scope.go:117] "RemoveContainer" containerID="76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1" Mar 12 18:10:29 crc kubenswrapper[4926]: E0312 18:10:29.470693 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1\": container with ID starting with 76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1 not found: ID does not exist" containerID="76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.470714 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1"} err="failed to get container status \"76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1\": rpc error: code = NotFound desc = could not find container \"76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1\": container with ID starting with 76aab1cc6cdb175ca5208c74f8c9cf48afd4b5c01686d1d1f1fa4ff001aa81c1 not found: ID does not exist" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.530705 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz6qt"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.531668 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/637236a6-6287-401d-a2cd-78713aa03176-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.536155 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wz6qt"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.549771 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c68kr"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.554694 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-c68kr"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.559407 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zx9h7"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.567107 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zx9h7"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.570590 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s4zf5"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.582771 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s4zf5"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.586874 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-565fl"] Mar 12 18:10:29 crc kubenswrapper[4926]: I0312 18:10:29.589839 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-565fl"] Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.226249 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" event={"ID":"daeebaaf-6a69-436e-b341-36fae756599e","Type":"ContainerStarted","Data":"e8e29782de63bdb9556d9ad371acf39d64a6ad349c1167e79b44e09709cbf79a"} Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.226602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" event={"ID":"daeebaaf-6a69-436e-b341-36fae756599e","Type":"ContainerStarted","Data":"cb5bce34c6486bdd5b618e103d0c85368e73bb247bb05dc441733b928f66d9e4"} Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.227638 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.232068 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.249922 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-d9gx5" podStartSLOduration=2.249908206 podStartE2EDuration="2.249908206s" podCreationTimestamp="2026-03-12 18:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:10:30.24779349 +0000 UTC m=+470.616419833" watchObservedRunningTime="2026-03-12 18:10:30.249908206 +0000 UTC m=+470.618534539" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481042 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-265z5"] Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481214 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481227 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481236 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481242 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481248 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481254 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481264 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481269 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481278 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481283 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481292 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481298 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481307 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481313 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481325 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481331 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481337 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481343 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481352 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481370 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="extract-content" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481378 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481384 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481392 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481397 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481405 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481411 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: E0312 18:10:30.481420 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481448 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="extract-utilities" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481552 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481561 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481569 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481579 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481587 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="637236a6-6287-401d-a2cd-78713aa03176" containerName="registry-server" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.481723 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" containerName="marketplace-operator" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.482189 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.493710 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.536140 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16fff1b6-a4ca-4ea9-aaba-40e9c136f62b" path="/var/lib/kubelet/pods/16fff1b6-a4ca-4ea9-aaba-40e9c136f62b/volumes" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.537859 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="637236a6-6287-401d-a2cd-78713aa03176" path="/var/lib/kubelet/pods/637236a6-6287-401d-a2cd-78713aa03176/volumes" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.539368 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="702daa2d-851e-4c3d-be86-4f337b4462f7" path="/var/lib/kubelet/pods/702daa2d-851e-4c3d-be86-4f337b4462f7/volumes" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.541751 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a609cd-c298-4356-9ddf-a7f125b52938" path="/var/lib/kubelet/pods/b2a609cd-c298-4356-9ddf-a7f125b52938/volumes" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.543625 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5fe4032-6a1e-4c27-9471-fa53e044826e" path="/var/lib/kubelet/pods/b5fe4032-6a1e-4c27-9471-fa53e044826e/volumes" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.545382 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-265z5"] Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.646062 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f746e64-bce3-4f58-b789-0f5573e28847-utilities\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.646226 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx687\" (UniqueName: \"kubernetes.io/projected/0f746e64-bce3-4f58-b789-0f5573e28847-kube-api-access-xx687\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.646345 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f746e64-bce3-4f58-b789-0f5573e28847-catalog-content\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.746976 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f746e64-bce3-4f58-b789-0f5573e28847-catalog-content\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.747347 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f746e64-bce3-4f58-b789-0f5573e28847-utilities\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.747513 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx687\" (UniqueName: \"kubernetes.io/projected/0f746e64-bce3-4f58-b789-0f5573e28847-kube-api-access-xx687\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.747525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f746e64-bce3-4f58-b789-0f5573e28847-catalog-content\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.748056 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f746e64-bce3-4f58-b789-0f5573e28847-utilities\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.775104 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx687\" (UniqueName: \"kubernetes.io/projected/0f746e64-bce3-4f58-b789-0f5573e28847-kube-api-access-xx687\") pod \"redhat-marketplace-265z5\" (UID: \"0f746e64-bce3-4f58-b789-0f5573e28847\") " pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:30 crc kubenswrapper[4926]: I0312 18:10:30.818027 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.008273 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-265z5"] Mar 12 18:10:31 crc kubenswrapper[4926]: W0312 18:10:31.016027 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f746e64_bce3_4f58_b789_0f5573e28847.slice/crio-27b0866546d967bb33c43963d5bf82a711ca8438fc5ee042557732aa8757e0a9 WatchSource:0}: Error finding container 27b0866546d967bb33c43963d5bf82a711ca8438fc5ee042557732aa8757e0a9: Status 404 returned error can't find the container with id 27b0866546d967bb33c43963d5bf82a711ca8438fc5ee042557732aa8757e0a9 Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.241858 4926 generic.go:334] "Generic (PLEG): container finished" podID="0f746e64-bce3-4f58-b789-0f5573e28847" containerID="527b0f3339ffe4707231354613a46abcba6e7c41812696c2338da4db5088147d" exitCode=0 Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.242091 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-265z5" event={"ID":"0f746e64-bce3-4f58-b789-0f5573e28847","Type":"ContainerDied","Data":"527b0f3339ffe4707231354613a46abcba6e7c41812696c2338da4db5088147d"} Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.242151 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-265z5" event={"ID":"0f746e64-bce3-4f58-b789-0f5573e28847","Type":"ContainerStarted","Data":"27b0866546d967bb33c43963d5bf82a711ca8438fc5ee042557732aa8757e0a9"} Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.479261 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-672mz"] Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.480952 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.483523 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.494506 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-672mz"] Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.661290 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-utilities\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.661360 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck6nj\" (UniqueName: \"kubernetes.io/projected/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-kube-api-access-ck6nj\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.661501 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-catalog-content\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.763185 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-utilities\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.763241 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck6nj\" (UniqueName: \"kubernetes.io/projected/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-kube-api-access-ck6nj\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.763304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-catalog-content\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.763787 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-catalog-content\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.763941 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-utilities\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.786869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck6nj\" (UniqueName: \"kubernetes.io/projected/c9f79c44-e93f-48ba-9f2d-8a5b61a86089-kube-api-access-ck6nj\") pod \"redhat-operators-672mz\" (UID: \"c9f79c44-e93f-48ba-9f2d-8a5b61a86089\") " pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:31 crc kubenswrapper[4926]: I0312 18:10:31.851824 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:32 crc kubenswrapper[4926]: I0312 18:10:32.088881 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-672mz"] Mar 12 18:10:32 crc kubenswrapper[4926]: I0312 18:10:32.248072 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-672mz" event={"ID":"c9f79c44-e93f-48ba-9f2d-8a5b61a86089","Type":"ContainerStarted","Data":"9d6815757925c31c05247e630d2a9254a2ea04f08098a0932fc5e7a8816ffd88"} Mar 12 18:10:32 crc kubenswrapper[4926]: I0312 18:10:32.248138 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-672mz" event={"ID":"c9f79c44-e93f-48ba-9f2d-8a5b61a86089","Type":"ContainerStarted","Data":"aa3a7f181e123af1866cec91223954325e07acf8dcad47201ae58efcffb9e372"} Mar 12 18:10:32 crc kubenswrapper[4926]: I0312 18:10:32.889012 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t8lwn"] Mar 12 18:10:32 crc kubenswrapper[4926]: I0312 18:10:32.894141 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:32 crc kubenswrapper[4926]: I0312 18:10:32.900926 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 18:10:32 crc kubenswrapper[4926]: I0312 18:10:32.902535 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t8lwn"] Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.027626 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ef972e3-81c7-4b62-aa07-4939aef86a2d-catalog-content\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.027687 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ef972e3-81c7-4b62-aa07-4939aef86a2d-utilities\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.027727 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f285p\" (UniqueName: \"kubernetes.io/projected/0ef972e3-81c7-4b62-aa07-4939aef86a2d-kube-api-access-f285p\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.128714 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ef972e3-81c7-4b62-aa07-4939aef86a2d-utilities\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.129152 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f285p\" (UniqueName: \"kubernetes.io/projected/0ef972e3-81c7-4b62-aa07-4939aef86a2d-kube-api-access-f285p\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.129380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ef972e3-81c7-4b62-aa07-4939aef86a2d-catalog-content\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.129509 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ef972e3-81c7-4b62-aa07-4939aef86a2d-utilities\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.129943 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ef972e3-81c7-4b62-aa07-4939aef86a2d-catalog-content\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.149693 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f285p\" (UniqueName: \"kubernetes.io/projected/0ef972e3-81c7-4b62-aa07-4939aef86a2d-kube-api-access-f285p\") pod \"certified-operators-t8lwn\" (UID: \"0ef972e3-81c7-4b62-aa07-4939aef86a2d\") " pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.226157 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.258131 4926 generic.go:334] "Generic (PLEG): container finished" podID="c9f79c44-e93f-48ba-9f2d-8a5b61a86089" containerID="9d6815757925c31c05247e630d2a9254a2ea04f08098a0932fc5e7a8816ffd88" exitCode=0 Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.258420 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-672mz" event={"ID":"c9f79c44-e93f-48ba-9f2d-8a5b61a86089","Type":"ContainerDied","Data":"9d6815757925c31c05247e630d2a9254a2ea04f08098a0932fc5e7a8816ffd88"} Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.264083 4926 generic.go:334] "Generic (PLEG): container finished" podID="0f746e64-bce3-4f58-b789-0f5573e28847" containerID="0508d641f408a89ec3f17b2ac05f9ede131af67d070a5d4c8ba77d8f2743a6d5" exitCode=0 Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.264147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-265z5" event={"ID":"0f746e64-bce3-4f58-b789-0f5573e28847","Type":"ContainerDied","Data":"0508d641f408a89ec3f17b2ac05f9ede131af67d070a5d4c8ba77d8f2743a6d5"} Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.456748 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t8lwn"] Mar 12 18:10:33 crc kubenswrapper[4926]: W0312 18:10:33.461214 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ef972e3_81c7_4b62_aa07_4939aef86a2d.slice/crio-fd2691aae6521c45040f9344f6d91ddad90e20cfd5d586ddeb2c5821d1336f0a WatchSource:0}: Error finding container fd2691aae6521c45040f9344f6d91ddad90e20cfd5d586ddeb2c5821d1336f0a: Status 404 returned error can't find the container with id fd2691aae6521c45040f9344f6d91ddad90e20cfd5d586ddeb2c5821d1336f0a Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.881006 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-89rth"] Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.883559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.885702 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.905379 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-89rth"] Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.952369 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-utilities\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.952485 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-catalog-content\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:33 crc kubenswrapper[4926]: I0312 18:10:33.952577 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5zfs\" (UniqueName: \"kubernetes.io/projected/8ed429f6-2923-42c2-a3b5-402c3dff4858-kube-api-access-c5zfs\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.053484 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-utilities\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.053538 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-catalog-content\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.053590 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5zfs\" (UniqueName: \"kubernetes.io/projected/8ed429f6-2923-42c2-a3b5-402c3dff4858-kube-api-access-c5zfs\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.054004 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-utilities\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.054101 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-catalog-content\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.084665 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5zfs\" (UniqueName: \"kubernetes.io/projected/8ed429f6-2923-42c2-a3b5-402c3dff4858-kube-api-access-c5zfs\") pod \"community-operators-89rth\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.227514 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.271639 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-672mz" event={"ID":"c9f79c44-e93f-48ba-9f2d-8a5b61a86089","Type":"ContainerStarted","Data":"7b03cdeb0a691f7942895be6b14554fa0f8ec4142145648b27f87e5596fa1020"} Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.273492 4926 generic.go:334] "Generic (PLEG): container finished" podID="0ef972e3-81c7-4b62-aa07-4939aef86a2d" containerID="cde3e1686aebfaaac97245815c4010a7e06082d4c42b0de785931985077be174" exitCode=0 Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.273651 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8lwn" event={"ID":"0ef972e3-81c7-4b62-aa07-4939aef86a2d","Type":"ContainerDied","Data":"cde3e1686aebfaaac97245815c4010a7e06082d4c42b0de785931985077be174"} Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.273969 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8lwn" event={"ID":"0ef972e3-81c7-4b62-aa07-4939aef86a2d","Type":"ContainerStarted","Data":"fd2691aae6521c45040f9344f6d91ddad90e20cfd5d586ddeb2c5821d1336f0a"} Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.276980 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-265z5" event={"ID":"0f746e64-bce3-4f58-b789-0f5573e28847","Type":"ContainerStarted","Data":"ed3558d1c2fe7aee056d002e16a5435263061c33364b778a5e945656d0caf11f"} Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.331627 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-265z5" podStartSLOduration=1.629136209 podStartE2EDuration="4.33160855s" podCreationTimestamp="2026-03-12 18:10:30 +0000 UTC" firstStartedPulling="2026-03-12 18:10:31.244385587 +0000 UTC m=+471.613011920" lastFinishedPulling="2026-03-12 18:10:33.946857918 +0000 UTC m=+474.315484261" observedRunningTime="2026-03-12 18:10:34.328866325 +0000 UTC m=+474.697492678" watchObservedRunningTime="2026-03-12 18:10:34.33160855 +0000 UTC m=+474.700234883" Mar 12 18:10:34 crc kubenswrapper[4926]: I0312 18:10:34.426790 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-89rth"] Mar 12 18:10:35 crc kubenswrapper[4926]: I0312 18:10:35.292096 4926 generic.go:334] "Generic (PLEG): container finished" podID="c9f79c44-e93f-48ba-9f2d-8a5b61a86089" containerID="7b03cdeb0a691f7942895be6b14554fa0f8ec4142145648b27f87e5596fa1020" exitCode=0 Mar 12 18:10:35 crc kubenswrapper[4926]: I0312 18:10:35.292265 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-672mz" event={"ID":"c9f79c44-e93f-48ba-9f2d-8a5b61a86089","Type":"ContainerDied","Data":"7b03cdeb0a691f7942895be6b14554fa0f8ec4142145648b27f87e5596fa1020"} Mar 12 18:10:35 crc kubenswrapper[4926]: I0312 18:10:35.310197 4926 generic.go:334] "Generic (PLEG): container finished" podID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerID="b4a2b0d2b334eaa88bb6c32b357510979e7382c6582dd687f2155b5f20b07067" exitCode=0 Mar 12 18:10:35 crc kubenswrapper[4926]: I0312 18:10:35.311181 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89rth" event={"ID":"8ed429f6-2923-42c2-a3b5-402c3dff4858","Type":"ContainerDied","Data":"b4a2b0d2b334eaa88bb6c32b357510979e7382c6582dd687f2155b5f20b07067"} Mar 12 18:10:35 crc kubenswrapper[4926]: I0312 18:10:35.311457 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89rth" event={"ID":"8ed429f6-2923-42c2-a3b5-402c3dff4858","Type":"ContainerStarted","Data":"75b5fa3c244e1a8120e460f27209d8d4bc13d27e6ea5e2efce3f0acb11f95356"} Mar 12 18:10:36 crc kubenswrapper[4926]: I0312 18:10:36.318542 4926 generic.go:334] "Generic (PLEG): container finished" podID="0ef972e3-81c7-4b62-aa07-4939aef86a2d" containerID="9986cda75fb5c1cbdafdcc2f593e2eca5328cbc3ead61fc97e07ee3b4410da7d" exitCode=0 Mar 12 18:10:36 crc kubenswrapper[4926]: I0312 18:10:36.318660 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8lwn" event={"ID":"0ef972e3-81c7-4b62-aa07-4939aef86a2d","Type":"ContainerDied","Data":"9986cda75fb5c1cbdafdcc2f593e2eca5328cbc3ead61fc97e07ee3b4410da7d"} Mar 12 18:10:36 crc kubenswrapper[4926]: I0312 18:10:36.323207 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-672mz" event={"ID":"c9f79c44-e93f-48ba-9f2d-8a5b61a86089","Type":"ContainerStarted","Data":"e5916d747c43eec52f6e9fdc5fce1e305ff6c9af8aaa450a2ea24c339f9fdd70"} Mar 12 18:10:36 crc kubenswrapper[4926]: I0312 18:10:36.356294 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-672mz" podStartSLOduration=2.785500238 podStartE2EDuration="5.356278331s" podCreationTimestamp="2026-03-12 18:10:31 +0000 UTC" firstStartedPulling="2026-03-12 18:10:33.259842591 +0000 UTC m=+473.628468924" lastFinishedPulling="2026-03-12 18:10:35.830620643 +0000 UTC m=+476.199247017" observedRunningTime="2026-03-12 18:10:36.355462695 +0000 UTC m=+476.724089038" watchObservedRunningTime="2026-03-12 18:10:36.356278331 +0000 UTC m=+476.724904664" Mar 12 18:10:37 crc kubenswrapper[4926]: I0312 18:10:37.330580 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t8lwn" event={"ID":"0ef972e3-81c7-4b62-aa07-4939aef86a2d","Type":"ContainerStarted","Data":"0b0384766172a95d6a3558b13f8ff146e41e28421d2d906ff78bcb5fc47029c5"} Mar 12 18:10:37 crc kubenswrapper[4926]: I0312 18:10:37.332949 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89rth" event={"ID":"8ed429f6-2923-42c2-a3b5-402c3dff4858","Type":"ContainerStarted","Data":"2efb7a52879f66525ffb82d5cb820bfe29543aedd6332a889b48b179b150edf8"} Mar 12 18:10:37 crc kubenswrapper[4926]: I0312 18:10:37.362865 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t8lwn" podStartSLOduration=2.519301066 podStartE2EDuration="5.362849807s" podCreationTimestamp="2026-03-12 18:10:32 +0000 UTC" firstStartedPulling="2026-03-12 18:10:34.274677983 +0000 UTC m=+474.643304316" lastFinishedPulling="2026-03-12 18:10:37.118226684 +0000 UTC m=+477.486853057" observedRunningTime="2026-03-12 18:10:37.362595789 +0000 UTC m=+477.731222132" watchObservedRunningTime="2026-03-12 18:10:37.362849807 +0000 UTC m=+477.731476140" Mar 12 18:10:38 crc kubenswrapper[4926]: I0312 18:10:38.341227 4926 generic.go:334] "Generic (PLEG): container finished" podID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerID="2efb7a52879f66525ffb82d5cb820bfe29543aedd6332a889b48b179b150edf8" exitCode=0 Mar 12 18:10:38 crc kubenswrapper[4926]: I0312 18:10:38.342830 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89rth" event={"ID":"8ed429f6-2923-42c2-a3b5-402c3dff4858","Type":"ContainerDied","Data":"2efb7a52879f66525ffb82d5cb820bfe29543aedd6332a889b48b179b150edf8"} Mar 12 18:10:39 crc kubenswrapper[4926]: I0312 18:10:39.349365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89rth" event={"ID":"8ed429f6-2923-42c2-a3b5-402c3dff4858","Type":"ContainerStarted","Data":"b6bb1201cf8704c1c5218ae5ad6eb53e93f35257e13f18d1b638bfe3b2da6322"} Mar 12 18:10:39 crc kubenswrapper[4926]: I0312 18:10:39.376197 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-89rth" podStartSLOduration=2.839829219 podStartE2EDuration="6.376175715s" podCreationTimestamp="2026-03-12 18:10:33 +0000 UTC" firstStartedPulling="2026-03-12 18:10:35.311691414 +0000 UTC m=+475.680317757" lastFinishedPulling="2026-03-12 18:10:38.84803791 +0000 UTC m=+479.216664253" observedRunningTime="2026-03-12 18:10:39.370575231 +0000 UTC m=+479.739201614" watchObservedRunningTime="2026-03-12 18:10:39.376175715 +0000 UTC m=+479.744802058" Mar 12 18:10:40 crc kubenswrapper[4926]: I0312 18:10:40.818193 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:40 crc kubenswrapper[4926]: I0312 18:10:40.818581 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:40 crc kubenswrapper[4926]: I0312 18:10:40.871808 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:41 crc kubenswrapper[4926]: I0312 18:10:41.405793 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-265z5" Mar 12 18:10:41 crc kubenswrapper[4926]: I0312 18:10:41.853040 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:41 crc kubenswrapper[4926]: I0312 18:10:41.853388 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:42 crc kubenswrapper[4926]: I0312 18:10:42.917878 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-672mz" podUID="c9f79c44-e93f-48ba-9f2d-8a5b61a86089" containerName="registry-server" probeResult="failure" output=< Mar 12 18:10:42 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:10:42 crc kubenswrapper[4926]: > Mar 12 18:10:43 crc kubenswrapper[4926]: I0312 18:10:43.227059 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:43 crc kubenswrapper[4926]: I0312 18:10:43.227119 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:43 crc kubenswrapper[4926]: I0312 18:10:43.281089 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:43 crc kubenswrapper[4926]: I0312 18:10:43.433259 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t8lwn" Mar 12 18:10:44 crc kubenswrapper[4926]: I0312 18:10:44.228253 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:44 crc kubenswrapper[4926]: I0312 18:10:44.229019 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:44 crc kubenswrapper[4926]: I0312 18:10:44.266862 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:44 crc kubenswrapper[4926]: I0312 18:10:44.417615 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.250477 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" podUID="97b0faa2-bcb2-417e-9065-3156860a8644" containerName="registry" containerID="cri-o://f7c989898f3b81f561b1ebddb62bcb7b02ee6fac8bd401ab3a5ff75b154fda11" gracePeriod=30 Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.387473 4926 generic.go:334] "Generic (PLEG): container finished" podID="97b0faa2-bcb2-417e-9065-3156860a8644" containerID="f7c989898f3b81f561b1ebddb62bcb7b02ee6fac8bd401ab3a5ff75b154fda11" exitCode=0 Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.388128 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" event={"ID":"97b0faa2-bcb2-417e-9065-3156860a8644","Type":"ContainerDied","Data":"f7c989898f3b81f561b1ebddb62bcb7b02ee6fac8bd401ab3a5ff75b154fda11"} Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.850765 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.942746 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-registry-certificates\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.943075 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltlt7\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-kube-api-access-ltlt7\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.944191 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-bound-sa-token\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.944363 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-registry-tls\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.944502 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97b0faa2-bcb2-417e-9065-3156860a8644-installation-pull-secrets\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.944749 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.944869 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-trusted-ca\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.944963 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97b0faa2-bcb2-417e-9065-3156860a8644-ca-trust-extracted\") pod \"97b0faa2-bcb2-417e-9065-3156860a8644\" (UID: \"97b0faa2-bcb2-417e-9065-3156860a8644\") " Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.943634 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.947839 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.956292 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.956399 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b0faa2-bcb2-417e-9065-3156860a8644-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.956564 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.960401 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.961548 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-kube-api-access-ltlt7" (OuterVolumeSpecName: "kube-api-access-ltlt7") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "kube-api-access-ltlt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:10:45 crc kubenswrapper[4926]: I0312 18:10:45.963800 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97b0faa2-bcb2-417e-9065-3156860a8644-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "97b0faa2-bcb2-417e-9065-3156860a8644" (UID: "97b0faa2-bcb2-417e-9065-3156860a8644"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.046486 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.046519 4926 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/97b0faa2-bcb2-417e-9065-3156860a8644-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.046532 4926 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/97b0faa2-bcb2-417e-9065-3156860a8644-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.046541 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltlt7\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-kube-api-access-ltlt7\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.046549 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.046558 4926 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/97b0faa2-bcb2-417e-9065-3156860a8644-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.046567 4926 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/97b0faa2-bcb2-417e-9065-3156860a8644-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.394674 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.398542 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g6fzt" event={"ID":"97b0faa2-bcb2-417e-9065-3156860a8644","Type":"ContainerDied","Data":"5fd53c65057b16b866539f523a9ef035c2f73d7900692a8369fc22f3d2f8f197"} Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.398637 4926 scope.go:117] "RemoveContainer" containerID="f7c989898f3b81f561b1ebddb62bcb7b02ee6fac8bd401ab3a5ff75b154fda11" Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.432996 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g6fzt"] Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.438132 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g6fzt"] Mar 12 18:10:46 crc kubenswrapper[4926]: I0312 18:10:46.497629 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97b0faa2-bcb2-417e-9065-3156860a8644" path="/var/lib/kubelet/pods/97b0faa2-bcb2-417e-9065-3156860a8644/volumes" Mar 12 18:10:51 crc kubenswrapper[4926]: I0312 18:10:51.921285 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:51 crc kubenswrapper[4926]: I0312 18:10:51.988710 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-672mz" Mar 12 18:10:56 crc kubenswrapper[4926]: I0312 18:10:56.817643 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:10:56 crc kubenswrapper[4926]: I0312 18:10:56.818207 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:10:56 crc kubenswrapper[4926]: I0312 18:10:56.818257 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:10:57 crc kubenswrapper[4926]: I0312 18:10:57.459039 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b3aa39c92eb410de660dbb6e5cf8c0dd506addf3c227d2622f913bd6b55014e2"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:10:57 crc kubenswrapper[4926]: I0312 18:10:57.459110 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://b3aa39c92eb410de660dbb6e5cf8c0dd506addf3c227d2622f913bd6b55014e2" gracePeriod=600 Mar 12 18:10:58 crc kubenswrapper[4926]: I0312 18:10:58.467464 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="b3aa39c92eb410de660dbb6e5cf8c0dd506addf3c227d2622f913bd6b55014e2" exitCode=0 Mar 12 18:10:58 crc kubenswrapper[4926]: I0312 18:10:58.467536 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"b3aa39c92eb410de660dbb6e5cf8c0dd506addf3c227d2622f913bd6b55014e2"} Mar 12 18:10:58 crc kubenswrapper[4926]: I0312 18:10:58.467900 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"869151c9e3071e8f72a54f977df4fbec55cdf81d3f75158a024a468d8d420c6b"} Mar 12 18:10:58 crc kubenswrapper[4926]: I0312 18:10:58.467919 4926 scope.go:117] "RemoveContainer" containerID="7d3bab13cabe4b82f90297599f822115d3fbb4c22873ec3b05761aca32e2caff" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.147734 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555652-mrcx4"] Mar 12 18:12:00 crc kubenswrapper[4926]: E0312 18:12:00.148518 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97b0faa2-bcb2-417e-9065-3156860a8644" containerName="registry" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.148533 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="97b0faa2-bcb2-417e-9065-3156860a8644" containerName="registry" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.148645 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="97b0faa2-bcb2-417e-9065-3156860a8644" containerName="registry" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.149032 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555652-mrcx4" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.152272 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.152508 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.156793 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.160965 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555652-mrcx4"] Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.164691 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsnjq\" (UniqueName: \"kubernetes.io/projected/a135018f-2c21-4678-a0fc-9d6b62dda2d6-kube-api-access-dsnjq\") pod \"auto-csr-approver-29555652-mrcx4\" (UID: \"a135018f-2c21-4678-a0fc-9d6b62dda2d6\") " pod="openshift-infra/auto-csr-approver-29555652-mrcx4" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.266384 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsnjq\" (UniqueName: \"kubernetes.io/projected/a135018f-2c21-4678-a0fc-9d6b62dda2d6-kube-api-access-dsnjq\") pod \"auto-csr-approver-29555652-mrcx4\" (UID: \"a135018f-2c21-4678-a0fc-9d6b62dda2d6\") " pod="openshift-infra/auto-csr-approver-29555652-mrcx4" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.292245 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsnjq\" (UniqueName: \"kubernetes.io/projected/a135018f-2c21-4678-a0fc-9d6b62dda2d6-kube-api-access-dsnjq\") pod \"auto-csr-approver-29555652-mrcx4\" (UID: \"a135018f-2c21-4678-a0fc-9d6b62dda2d6\") " pod="openshift-infra/auto-csr-approver-29555652-mrcx4" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.476955 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555652-mrcx4" Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.731016 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555652-mrcx4"] Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.743641 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:12:00 crc kubenswrapper[4926]: I0312 18:12:00.892897 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555652-mrcx4" event={"ID":"a135018f-2c21-4678-a0fc-9d6b62dda2d6","Type":"ContainerStarted","Data":"cd090b4c4adefae3d508f3c966780859889876275ce95ed1f8a017bfc8d2b803"} Mar 12 18:12:02 crc kubenswrapper[4926]: I0312 18:12:02.910809 4926 generic.go:334] "Generic (PLEG): container finished" podID="a135018f-2c21-4678-a0fc-9d6b62dda2d6" containerID="80892424cda12377cff2432a54ca031087805d38670f04b3a339fc6f8678512c" exitCode=0 Mar 12 18:12:02 crc kubenswrapper[4926]: I0312 18:12:02.910878 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555652-mrcx4" event={"ID":"a135018f-2c21-4678-a0fc-9d6b62dda2d6","Type":"ContainerDied","Data":"80892424cda12377cff2432a54ca031087805d38670f04b3a339fc6f8678512c"} Mar 12 18:12:04 crc kubenswrapper[4926]: I0312 18:12:04.284496 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555652-mrcx4" Mar 12 18:12:04 crc kubenswrapper[4926]: I0312 18:12:04.431285 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsnjq\" (UniqueName: \"kubernetes.io/projected/a135018f-2c21-4678-a0fc-9d6b62dda2d6-kube-api-access-dsnjq\") pod \"a135018f-2c21-4678-a0fc-9d6b62dda2d6\" (UID: \"a135018f-2c21-4678-a0fc-9d6b62dda2d6\") " Mar 12 18:12:04 crc kubenswrapper[4926]: I0312 18:12:04.439742 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a135018f-2c21-4678-a0fc-9d6b62dda2d6-kube-api-access-dsnjq" (OuterVolumeSpecName: "kube-api-access-dsnjq") pod "a135018f-2c21-4678-a0fc-9d6b62dda2d6" (UID: "a135018f-2c21-4678-a0fc-9d6b62dda2d6"). InnerVolumeSpecName "kube-api-access-dsnjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:12:04 crc kubenswrapper[4926]: I0312 18:12:04.533428 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsnjq\" (UniqueName: \"kubernetes.io/projected/a135018f-2c21-4678-a0fc-9d6b62dda2d6-kube-api-access-dsnjq\") on node \"crc\" DevicePath \"\"" Mar 12 18:12:04 crc kubenswrapper[4926]: I0312 18:12:04.927215 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555652-mrcx4" event={"ID":"a135018f-2c21-4678-a0fc-9d6b62dda2d6","Type":"ContainerDied","Data":"cd090b4c4adefae3d508f3c966780859889876275ce95ed1f8a017bfc8d2b803"} Mar 12 18:12:04 crc kubenswrapper[4926]: I0312 18:12:04.927272 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd090b4c4adefae3d508f3c966780859889876275ce95ed1f8a017bfc8d2b803" Mar 12 18:12:04 crc kubenswrapper[4926]: I0312 18:12:04.927296 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555652-mrcx4" Mar 12 18:12:05 crc kubenswrapper[4926]: I0312 18:12:05.342283 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555646-wqpkb"] Mar 12 18:12:05 crc kubenswrapper[4926]: I0312 18:12:05.346491 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555646-wqpkb"] Mar 12 18:12:06 crc kubenswrapper[4926]: I0312 18:12:06.499644 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d68160cf-4e6c-4294-bfdc-4acb74637ecb" path="/var/lib/kubelet/pods/d68160cf-4e6c-4294-bfdc-4acb74637ecb/volumes" Mar 12 18:13:26 crc kubenswrapper[4926]: I0312 18:13:26.818121 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:13:26 crc kubenswrapper[4926]: I0312 18:13:26.819008 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:13:56 crc kubenswrapper[4926]: I0312 18:13:56.817939 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:13:56 crc kubenswrapper[4926]: I0312 18:13:56.818787 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.142260 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555654-jjvcf"] Mar 12 18:14:00 crc kubenswrapper[4926]: E0312 18:14:00.142838 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a135018f-2c21-4678-a0fc-9d6b62dda2d6" containerName="oc" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.142856 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a135018f-2c21-4678-a0fc-9d6b62dda2d6" containerName="oc" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.142970 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a135018f-2c21-4678-a0fc-9d6b62dda2d6" containerName="oc" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.143411 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555654-jjvcf" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.146564 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.147222 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.147345 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.151297 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555654-jjvcf"] Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.316121 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsrbz\" (UniqueName: \"kubernetes.io/projected/b3e47fc3-fa77-4992-b1f8-dfff8b4d924e-kube-api-access-bsrbz\") pod \"auto-csr-approver-29555654-jjvcf\" (UID: \"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e\") " pod="openshift-infra/auto-csr-approver-29555654-jjvcf" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.417247 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsrbz\" (UniqueName: \"kubernetes.io/projected/b3e47fc3-fa77-4992-b1f8-dfff8b4d924e-kube-api-access-bsrbz\") pod \"auto-csr-approver-29555654-jjvcf\" (UID: \"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e\") " pod="openshift-infra/auto-csr-approver-29555654-jjvcf" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.450837 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsrbz\" (UniqueName: \"kubernetes.io/projected/b3e47fc3-fa77-4992-b1f8-dfff8b4d924e-kube-api-access-bsrbz\") pod \"auto-csr-approver-29555654-jjvcf\" (UID: \"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e\") " pod="openshift-infra/auto-csr-approver-29555654-jjvcf" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.466323 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555654-jjvcf" Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.754162 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555654-jjvcf"] Mar 12 18:14:00 crc kubenswrapper[4926]: I0312 18:14:00.956549 4926 scope.go:117] "RemoveContainer" containerID="52e8bc12e7903e1d736ebf550d285fa84f2e22f28d746c1ea7750000e16f3903" Mar 12 18:14:01 crc kubenswrapper[4926]: I0312 18:14:01.715539 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555654-jjvcf" event={"ID":"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e","Type":"ContainerStarted","Data":"89f51dfda28111729ff6b326d4927dd9f5c3592cb11d933b305b0ad0af84d6b4"} Mar 12 18:14:02 crc kubenswrapper[4926]: I0312 18:14:02.733267 4926 generic.go:334] "Generic (PLEG): container finished" podID="b3e47fc3-fa77-4992-b1f8-dfff8b4d924e" containerID="5d29d1a82fd23f75a9296b7891f008517b3e7945af33e9449e2dfcb52711dd8e" exitCode=0 Mar 12 18:14:02 crc kubenswrapper[4926]: I0312 18:14:02.733347 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555654-jjvcf" event={"ID":"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e","Type":"ContainerDied","Data":"5d29d1a82fd23f75a9296b7891f008517b3e7945af33e9449e2dfcb52711dd8e"} Mar 12 18:14:03 crc kubenswrapper[4926]: I0312 18:14:03.998196 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555654-jjvcf" Mar 12 18:14:04 crc kubenswrapper[4926]: I0312 18:14:04.172818 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsrbz\" (UniqueName: \"kubernetes.io/projected/b3e47fc3-fa77-4992-b1f8-dfff8b4d924e-kube-api-access-bsrbz\") pod \"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e\" (UID: \"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e\") " Mar 12 18:14:04 crc kubenswrapper[4926]: I0312 18:14:04.182957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3e47fc3-fa77-4992-b1f8-dfff8b4d924e-kube-api-access-bsrbz" (OuterVolumeSpecName: "kube-api-access-bsrbz") pod "b3e47fc3-fa77-4992-b1f8-dfff8b4d924e" (UID: "b3e47fc3-fa77-4992-b1f8-dfff8b4d924e"). InnerVolumeSpecName "kube-api-access-bsrbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:14:04 crc kubenswrapper[4926]: I0312 18:14:04.275137 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsrbz\" (UniqueName: \"kubernetes.io/projected/b3e47fc3-fa77-4992-b1f8-dfff8b4d924e-kube-api-access-bsrbz\") on node \"crc\" DevicePath \"\"" Mar 12 18:14:04 crc kubenswrapper[4926]: I0312 18:14:04.751027 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555654-jjvcf" event={"ID":"b3e47fc3-fa77-4992-b1f8-dfff8b4d924e","Type":"ContainerDied","Data":"89f51dfda28111729ff6b326d4927dd9f5c3592cb11d933b305b0ad0af84d6b4"} Mar 12 18:14:04 crc kubenswrapper[4926]: I0312 18:14:04.751128 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89f51dfda28111729ff6b326d4927dd9f5c3592cb11d933b305b0ad0af84d6b4" Mar 12 18:14:04 crc kubenswrapper[4926]: I0312 18:14:04.751137 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555654-jjvcf" Mar 12 18:14:05 crc kubenswrapper[4926]: I0312 18:14:05.075753 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555648-j9n25"] Mar 12 18:14:05 crc kubenswrapper[4926]: I0312 18:14:05.082053 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555648-j9n25"] Mar 12 18:14:06 crc kubenswrapper[4926]: I0312 18:14:06.505695 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dbfae2a-83a6-4586-a33c-b9bcfd4df092" path="/var/lib/kubelet/pods/8dbfae2a-83a6-4586-a33c-b9bcfd4df092/volumes" Mar 12 18:14:26 crc kubenswrapper[4926]: I0312 18:14:26.817274 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:14:26 crc kubenswrapper[4926]: I0312 18:14:26.817890 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:14:26 crc kubenswrapper[4926]: I0312 18:14:26.817965 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:14:26 crc kubenswrapper[4926]: I0312 18:14:26.818836 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"869151c9e3071e8f72a54f977df4fbec55cdf81d3f75158a024a468d8d420c6b"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:14:26 crc kubenswrapper[4926]: I0312 18:14:26.818948 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://869151c9e3071e8f72a54f977df4fbec55cdf81d3f75158a024a468d8d420c6b" gracePeriod=600 Mar 12 18:14:27 crc kubenswrapper[4926]: I0312 18:14:27.924349 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="869151c9e3071e8f72a54f977df4fbec55cdf81d3f75158a024a468d8d420c6b" exitCode=0 Mar 12 18:14:27 crc kubenswrapper[4926]: I0312 18:14:27.924508 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"869151c9e3071e8f72a54f977df4fbec55cdf81d3f75158a024a468d8d420c6b"} Mar 12 18:14:27 crc kubenswrapper[4926]: I0312 18:14:27.925140 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"a397bef079b1410b3294983dad25ada9109b1a0eac364c78c0ff4aeeccdf38ed"} Mar 12 18:14:27 crc kubenswrapper[4926]: I0312 18:14:27.925215 4926 scope.go:117] "RemoveContainer" containerID="b3aa39c92eb410de660dbb6e5cf8c0dd506addf3c227d2622f913bd6b55014e2" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.159061 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6"] Mar 12 18:15:00 crc kubenswrapper[4926]: E0312 18:15:00.160575 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3e47fc3-fa77-4992-b1f8-dfff8b4d924e" containerName="oc" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.160608 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3e47fc3-fa77-4992-b1f8-dfff8b4d924e" containerName="oc" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.160881 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3e47fc3-fa77-4992-b1f8-dfff8b4d924e" containerName="oc" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.161868 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.168948 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.174100 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.176080 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6"] Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.257955 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69533265-f2d7-4566-a876-732db938b254-config-volume\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.258038 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69533265-f2d7-4566-a876-732db938b254-secret-volume\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.258253 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krf5x\" (UniqueName: \"kubernetes.io/projected/69533265-f2d7-4566-a876-732db938b254-kube-api-access-krf5x\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.359837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69533265-f2d7-4566-a876-732db938b254-secret-volume\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.359984 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krf5x\" (UniqueName: \"kubernetes.io/projected/69533265-f2d7-4566-a876-732db938b254-kube-api-access-krf5x\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.360050 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69533265-f2d7-4566-a876-732db938b254-config-volume\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.361499 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69533265-f2d7-4566-a876-732db938b254-config-volume\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.368778 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69533265-f2d7-4566-a876-732db938b254-secret-volume\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.390029 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krf5x\" (UniqueName: \"kubernetes.io/projected/69533265-f2d7-4566-a876-732db938b254-kube-api-access-krf5x\") pod \"collect-profiles-29555655-pzwh6\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.495581 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:00 crc kubenswrapper[4926]: I0312 18:15:00.724829 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6"] Mar 12 18:15:01 crc kubenswrapper[4926]: I0312 18:15:01.023968 4926 scope.go:117] "RemoveContainer" containerID="57bd95669fead78f455d353e126bbd0157addbc0b0171c28ad86c1bdee789263" Mar 12 18:15:01 crc kubenswrapper[4926]: I0312 18:15:01.167520 4926 generic.go:334] "Generic (PLEG): container finished" podID="69533265-f2d7-4566-a876-732db938b254" containerID="faccc149e57f019c53b37ac265b109708c4a9cb3385ebb8ecda1f786761d99fb" exitCode=0 Mar 12 18:15:01 crc kubenswrapper[4926]: I0312 18:15:01.167566 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" event={"ID":"69533265-f2d7-4566-a876-732db938b254","Type":"ContainerDied","Data":"faccc149e57f019c53b37ac265b109708c4a9cb3385ebb8ecda1f786761d99fb"} Mar 12 18:15:01 crc kubenswrapper[4926]: I0312 18:15:01.167600 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" event={"ID":"69533265-f2d7-4566-a876-732db938b254","Type":"ContainerStarted","Data":"c122a8e907362d7f0e0eb0123d6829139638732723b911d7f545569fdcbc25b6"} Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.467353 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.490373 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krf5x\" (UniqueName: \"kubernetes.io/projected/69533265-f2d7-4566-a876-732db938b254-kube-api-access-krf5x\") pod \"69533265-f2d7-4566-a876-732db938b254\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.490427 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69533265-f2d7-4566-a876-732db938b254-config-volume\") pod \"69533265-f2d7-4566-a876-732db938b254\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.490526 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69533265-f2d7-4566-a876-732db938b254-secret-volume\") pod \"69533265-f2d7-4566-a876-732db938b254\" (UID: \"69533265-f2d7-4566-a876-732db938b254\") " Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.491288 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69533265-f2d7-4566-a876-732db938b254-config-volume" (OuterVolumeSpecName: "config-volume") pod "69533265-f2d7-4566-a876-732db938b254" (UID: "69533265-f2d7-4566-a876-732db938b254"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.497645 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69533265-f2d7-4566-a876-732db938b254-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "69533265-f2d7-4566-a876-732db938b254" (UID: "69533265-f2d7-4566-a876-732db938b254"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.499112 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69533265-f2d7-4566-a876-732db938b254-kube-api-access-krf5x" (OuterVolumeSpecName: "kube-api-access-krf5x") pod "69533265-f2d7-4566-a876-732db938b254" (UID: "69533265-f2d7-4566-a876-732db938b254"). InnerVolumeSpecName "kube-api-access-krf5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.592223 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69533265-f2d7-4566-a876-732db938b254-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.592267 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krf5x\" (UniqueName: \"kubernetes.io/projected/69533265-f2d7-4566-a876-732db938b254-kube-api-access-krf5x\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:02 crc kubenswrapper[4926]: I0312 18:15:02.592280 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69533265-f2d7-4566-a876-732db938b254-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:03 crc kubenswrapper[4926]: I0312 18:15:03.198801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" event={"ID":"69533265-f2d7-4566-a876-732db938b254","Type":"ContainerDied","Data":"c122a8e907362d7f0e0eb0123d6829139638732723b911d7f545569fdcbc25b6"} Mar 12 18:15:03 crc kubenswrapper[4926]: I0312 18:15:03.199298 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c122a8e907362d7f0e0eb0123d6829139638732723b911d7f545569fdcbc25b6" Mar 12 18:15:03 crc kubenswrapper[4926]: I0312 18:15:03.198884 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555655-pzwh6" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.282497 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc"] Mar 12 18:15:43 crc kubenswrapper[4926]: E0312 18:15:43.283197 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69533265-f2d7-4566-a876-732db938b254" containerName="collect-profiles" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.283211 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="69533265-f2d7-4566-a876-732db938b254" containerName="collect-profiles" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.283325 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="69533265-f2d7-4566-a876-732db938b254" containerName="collect-profiles" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.285988 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.292506 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-spdvk" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.292643 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.292758 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.295084 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc"] Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.303746 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-7pf7h"] Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.304531 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-7pf7h" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.321004 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-dqkb8" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.352792 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5kmk\" (UniqueName: \"kubernetes.io/projected/04933adf-efe6-4d54-8575-cc5c4069ea9a-kube-api-access-n5kmk\") pod \"cert-manager-858654f9db-7pf7h\" (UID: \"04933adf-efe6-4d54-8575-cc5c4069ea9a\") " pod="cert-manager/cert-manager-858654f9db-7pf7h" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.353129 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzx22\" (UniqueName: \"kubernetes.io/projected/eaa86db1-fe85-4b00-b8e0-c61cb013f52d-kube-api-access-pzx22\") pod \"cert-manager-cainjector-cf98fcc89-mbhfc\" (UID: \"eaa86db1-fe85-4b00-b8e0-c61cb013f52d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.355412 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-p8p7c"] Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.358131 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.359966 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-tdvf8" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.365016 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-7pf7h"] Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.368928 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-p8p7c"] Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.453855 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzx22\" (UniqueName: \"kubernetes.io/projected/eaa86db1-fe85-4b00-b8e0-c61cb013f52d-kube-api-access-pzx22\") pod \"cert-manager-cainjector-cf98fcc89-mbhfc\" (UID: \"eaa86db1-fe85-4b00-b8e0-c61cb013f52d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.453944 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5kmk\" (UniqueName: \"kubernetes.io/projected/04933adf-efe6-4d54-8575-cc5c4069ea9a-kube-api-access-n5kmk\") pod \"cert-manager-858654f9db-7pf7h\" (UID: \"04933adf-efe6-4d54-8575-cc5c4069ea9a\") " pod="cert-manager/cert-manager-858654f9db-7pf7h" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.477447 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5kmk\" (UniqueName: \"kubernetes.io/projected/04933adf-efe6-4d54-8575-cc5c4069ea9a-kube-api-access-n5kmk\") pod \"cert-manager-858654f9db-7pf7h\" (UID: \"04933adf-efe6-4d54-8575-cc5c4069ea9a\") " pod="cert-manager/cert-manager-858654f9db-7pf7h" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.481624 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzx22\" (UniqueName: \"kubernetes.io/projected/eaa86db1-fe85-4b00-b8e0-c61cb013f52d-kube-api-access-pzx22\") pod \"cert-manager-cainjector-cf98fcc89-mbhfc\" (UID: \"eaa86db1-fe85-4b00-b8e0-c61cb013f52d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.555138 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86966\" (UniqueName: \"kubernetes.io/projected/ef14eb59-d30a-437c-80d1-70513a544b2d-kube-api-access-86966\") pod \"cert-manager-webhook-687f57d79b-p8p7c\" (UID: \"ef14eb59-d30a-437c-80d1-70513a544b2d\") " pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.605685 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.647267 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-7pf7h" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.656726 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86966\" (UniqueName: \"kubernetes.io/projected/ef14eb59-d30a-437c-80d1-70513a544b2d-kube-api-access-86966\") pod \"cert-manager-webhook-687f57d79b-p8p7c\" (UID: \"ef14eb59-d30a-437c-80d1-70513a544b2d\") " pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.680154 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86966\" (UniqueName: \"kubernetes.io/projected/ef14eb59-d30a-437c-80d1-70513a544b2d-kube-api-access-86966\") pod \"cert-manager-webhook-687f57d79b-p8p7c\" (UID: \"ef14eb59-d30a-437c-80d1-70513a544b2d\") " pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.841394 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc"] Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.889655 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-7pf7h"] Mar 12 18:15:43 crc kubenswrapper[4926]: W0312 18:15:43.895524 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04933adf_efe6_4d54_8575_cc5c4069ea9a.slice/crio-7d3bd66ee4d91c13c77673eccd2270176a8715c827c63ca1ddcfab990f1f37f1 WatchSource:0}: Error finding container 7d3bd66ee4d91c13c77673eccd2270176a8715c827c63ca1ddcfab990f1f37f1: Status 404 returned error can't find the container with id 7d3bd66ee4d91c13c77673eccd2270176a8715c827c63ca1ddcfab990f1f37f1 Mar 12 18:15:43 crc kubenswrapper[4926]: I0312 18:15:43.974149 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" Mar 12 18:15:44 crc kubenswrapper[4926]: I0312 18:15:44.173904 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-p8p7c"] Mar 12 18:15:44 crc kubenswrapper[4926]: W0312 18:15:44.180632 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef14eb59_d30a_437c_80d1_70513a544b2d.slice/crio-820bc38aea3d9c60d3d68cfc4b4fb00ab5ec20a971c76a654f49c505bd143284 WatchSource:0}: Error finding container 820bc38aea3d9c60d3d68cfc4b4fb00ab5ec20a971c76a654f49c505bd143284: Status 404 returned error can't find the container with id 820bc38aea3d9c60d3d68cfc4b4fb00ab5ec20a971c76a654f49c505bd143284 Mar 12 18:15:44 crc kubenswrapper[4926]: I0312 18:15:44.506799 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" event={"ID":"ef14eb59-d30a-437c-80d1-70513a544b2d","Type":"ContainerStarted","Data":"820bc38aea3d9c60d3d68cfc4b4fb00ab5ec20a971c76a654f49c505bd143284"} Mar 12 18:15:44 crc kubenswrapper[4926]: I0312 18:15:44.506878 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" event={"ID":"eaa86db1-fe85-4b00-b8e0-c61cb013f52d","Type":"ContainerStarted","Data":"2497c20f07fed49993dc459fd6d32fb2a0a69d28ec2ce1932b80d92a25e239f2"} Mar 12 18:15:44 crc kubenswrapper[4926]: I0312 18:15:44.507106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-7pf7h" event={"ID":"04933adf-efe6-4d54-8575-cc5c4069ea9a","Type":"ContainerStarted","Data":"7d3bd66ee4d91c13c77673eccd2270176a8715c827c63ca1ddcfab990f1f37f1"} Mar 12 18:15:48 crc kubenswrapper[4926]: I0312 18:15:48.533686 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-7pf7h" event={"ID":"04933adf-efe6-4d54-8575-cc5c4069ea9a","Type":"ContainerStarted","Data":"75169479de569e14a862d459dfadcad2bbda4900c5eadc77a05da06117ce0055"} Mar 12 18:15:48 crc kubenswrapper[4926]: I0312 18:15:48.536339 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" event={"ID":"ef14eb59-d30a-437c-80d1-70513a544b2d","Type":"ContainerStarted","Data":"e2e67a1a1d0b57520b59b6c5c98db19e3f96b7bcd647c2b70927945105dc4008"} Mar 12 18:15:48 crc kubenswrapper[4926]: I0312 18:15:48.536554 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" Mar 12 18:15:48 crc kubenswrapper[4926]: I0312 18:15:48.538965 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" event={"ID":"eaa86db1-fe85-4b00-b8e0-c61cb013f52d","Type":"ContainerStarted","Data":"70cd182f90e9394d2f17d256e0e0e5c9b125bc76c12c13592271a70812214135"} Mar 12 18:15:48 crc kubenswrapper[4926]: I0312 18:15:48.564718 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-7pf7h" podStartSLOduration=1.7459865749999999 podStartE2EDuration="5.564690053s" podCreationTimestamp="2026-03-12 18:15:43 +0000 UTC" firstStartedPulling="2026-03-12 18:15:43.897720114 +0000 UTC m=+784.266346447" lastFinishedPulling="2026-03-12 18:15:47.716423592 +0000 UTC m=+788.085049925" observedRunningTime="2026-03-12 18:15:48.557596392 +0000 UTC m=+788.926222785" watchObservedRunningTime="2026-03-12 18:15:48.564690053 +0000 UTC m=+788.933316426" Mar 12 18:15:48 crc kubenswrapper[4926]: I0312 18:15:48.577923 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" podStartSLOduration=2.104374805 podStartE2EDuration="5.577898694s" podCreationTimestamp="2026-03-12 18:15:43 +0000 UTC" firstStartedPulling="2026-03-12 18:15:44.184234647 +0000 UTC m=+784.552860980" lastFinishedPulling="2026-03-12 18:15:47.657758536 +0000 UTC m=+788.026384869" observedRunningTime="2026-03-12 18:15:48.574238589 +0000 UTC m=+788.942864992" watchObservedRunningTime="2026-03-12 18:15:48.577898694 +0000 UTC m=+788.946525057" Mar 12 18:15:48 crc kubenswrapper[4926]: I0312 18:15:48.590299 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-mbhfc" podStartSLOduration=1.78312085 podStartE2EDuration="5.590275669s" podCreationTimestamp="2026-03-12 18:15:43 +0000 UTC" firstStartedPulling="2026-03-12 18:15:43.851144574 +0000 UTC m=+784.219770907" lastFinishedPulling="2026-03-12 18:15:47.658299393 +0000 UTC m=+788.026925726" observedRunningTime="2026-03-12 18:15:48.587865264 +0000 UTC m=+788.956491627" watchObservedRunningTime="2026-03-12 18:15:48.590275669 +0000 UTC m=+788.958902042" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.468586 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zlfmg"] Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.469977 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-controller" containerID="cri-o://c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.470063 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="nbdb" containerID="cri-o://c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.470128 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.470196 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-node" containerID="cri-o://d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.470178 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="sbdb" containerID="cri-o://0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.470294 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="northd" containerID="cri-o://88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.470304 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-acl-logging" containerID="cri-o://1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.534011 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" containerID="cri-o://a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" gracePeriod=30 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.620342 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/3.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.630614 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovn-acl-logging/0.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.631080 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovn-controller/0.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.631419 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" exitCode=0 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.631476 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" exitCode=143 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.631524 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.631554 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.634031 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/2.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.635059 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/1.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.635104 4926 generic.go:334] "Generic (PLEG): container finished" podID="d5a53ef4-c701-457f-9cf2-85819bf04d1a" containerID="2c510c831017eeb7aed88601040d4790cf6d5bebce7b09f246c92ea9b81e2481" exitCode=2 Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.635141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerDied","Data":"2c510c831017eeb7aed88601040d4790cf6d5bebce7b09f246c92ea9b81e2481"} Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.635178 4926 scope.go:117] "RemoveContainer" containerID="bb6d6fe627e09e41640be0175c96b2d983a5b9f7b7e50c1792cfda71adaf2cf4" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.635697 4926 scope.go:117] "RemoveContainer" containerID="2c510c831017eeb7aed88601040d4790cf6d5bebce7b09f246c92ea9b81e2481" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.635967 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xwqvl_openshift-multus(d5a53ef4-c701-457f-9cf2-85819bf04d1a)\"" pod="openshift-multus/multus-xwqvl" podUID="d5a53ef4-c701-457f-9cf2-85819bf04d1a" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.761327 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/3.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.764214 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovn-acl-logging/0.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.764787 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovn-controller/0.log" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.765264 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.832426 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-94d7p"] Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.832775 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.832804 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.832823 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.832836 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.832879 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-node" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.832894 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-node" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.832911 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="nbdb" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.832922 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="nbdb" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.832944 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="northd" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.832956 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="northd" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.832974 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-acl-logging" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.832987 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-acl-logging" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.833006 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833020 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.833035 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kubecfg-setup" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833047 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kubecfg-setup" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.833065 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="sbdb" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833076 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="sbdb" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.833097 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833109 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.833126 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833138 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833304 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="northd" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833319 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833334 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-node" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833352 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="sbdb" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833365 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-acl-logging" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833379 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833396 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833415 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833429 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833493 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="nbdb" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833508 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovn-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.833686 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833701 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: E0312 18:15:53.833718 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833731 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.833910 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerName="ovnkube-controller" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.836947 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916509 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovn-node-metrics-cert\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916586 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-script-lib\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916619 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-var-lib-openvswitch\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916646 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-kubelet\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916667 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-ovn-kubernetes\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916694 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t5dd\" (UniqueName: \"kubernetes.io/projected/bc33af41-5aa0-4254-ac75-69433d5f4ce9-kube-api-access-4t5dd\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916718 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-openvswitch\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916736 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-ovn\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916757 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-log-socket\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916797 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916816 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-node-log\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916834 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-etc-openvswitch\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-slash\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916885 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-env-overrides\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916904 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-systemd-units\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916939 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-bin\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.916984 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-config\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917005 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-netd\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917028 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-systemd\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917046 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-netns\") pod \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\" (UID: \"bc33af41-5aa0-4254-ac75-69433d5f4ce9\") " Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917283 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917322 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-node-log" (OuterVolumeSpecName: "node-log") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917345 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917344 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917366 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-slash" (OuterVolumeSpecName: "host-slash") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917775 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917832 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.917856 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918080 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918115 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918133 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918150 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918149 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918167 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918180 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918236 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.918282 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-log-socket" (OuterVolumeSpecName: "log-socket") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.925104 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.925120 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc33af41-5aa0-4254-ac75-69433d5f4ce9-kube-api-access-4t5dd" (OuterVolumeSpecName: "kube-api-access-4t5dd") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "kube-api-access-4t5dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.932273 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "bc33af41-5aa0-4254-ac75-69433d5f4ce9" (UID: "bc33af41-5aa0-4254-ac75-69433d5f4ce9"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:15:53 crc kubenswrapper[4926]: I0312 18:15:53.977648 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-p8p7c" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018429 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-etc-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018524 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-systemd\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018576 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-cni-bin\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018615 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-slash\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018659 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-cni-netd\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018695 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-systemd-units\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018820 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-env-overrides\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.018994 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovn-node-metrics-cert\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019078 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019106 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-node-log\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019168 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-var-lib-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019249 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovnkube-config\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019265 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-kubelet\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019292 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-run-ovn-kubernetes\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019324 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llt9q\" (UniqueName: \"kubernetes.io/projected/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-kube-api-access-llt9q\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019343 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-run-netns\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019744 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-ovn\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019815 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-log-socket\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.019871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovnkube-script-lib\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020075 4926 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020108 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020131 4926 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020154 4926 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020176 4926 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020200 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020229 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020255 4926 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020280 4926 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020304 4926 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020322 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t5dd\" (UniqueName: \"kubernetes.io/projected/bc33af41-5aa0-4254-ac75-69433d5f4ce9-kube-api-access-4t5dd\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020341 4926 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020359 4926 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020380 4926 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-log-socket\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020400 4926 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020420 4926 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-node-log\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020481 4926 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020500 4926 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-host-slash\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020519 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc33af41-5aa0-4254-ac75-69433d5f4ce9-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.020538 4926 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc33af41-5aa0-4254-ac75-69433d5f4ce9-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121500 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-cni-bin\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121561 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-slash\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121611 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-cni-netd\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121652 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-systemd-units\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121684 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-env-overrides\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121721 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121720 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-slash\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121752 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovn-node-metrics-cert\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121779 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-cni-netd\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121920 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121927 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-systemd-units\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121973 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121965 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-node-log\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.121939 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-cni-bin\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122018 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-node-log\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122094 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-var-lib-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122170 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovnkube-config\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122202 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-kubelet\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122226 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-var-lib-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-run-netns\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-run-netns\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122284 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-run-ovn-kubernetes\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122314 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-kubelet\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122318 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llt9q\" (UniqueName: \"kubernetes.io/projected/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-kube-api-access-llt9q\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122390 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-ovn\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122435 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-log-socket\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122509 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovnkube-script-lib\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122571 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-etc-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122605 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-systemd\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122742 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-systemd\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122756 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-host-run-ovn-kubernetes\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122785 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-run-ovn\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122843 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-etc-openvswitch\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.122931 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-log-socket\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.123168 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovnkube-config\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.123514 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-env-overrides\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.123990 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovnkube-script-lib\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.126617 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-ovn-node-metrics-cert\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.152585 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llt9q\" (UniqueName: \"kubernetes.io/projected/2054c0e9-f3e6-4d68-ab8f-57466f8b3e47-kube-api-access-llt9q\") pod \"ovnkube-node-94d7p\" (UID: \"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47\") " pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.163409 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.644289 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/2.log" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.646759 4926 generic.go:334] "Generic (PLEG): container finished" podID="2054c0e9-f3e6-4d68-ab8f-57466f8b3e47" containerID="0017057d393b7fac29161022acddc46b04957b0c846ef899e05ba04301738f66" exitCode=0 Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.646892 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerDied","Data":"0017057d393b7fac29161022acddc46b04957b0c846ef899e05ba04301738f66"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.646996 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"33c8027e2c14712760adfead899aee80df4c4ebc8bcac548b0a119c466eef8a2"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.650805 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovnkube-controller/3.log" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.656770 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovn-acl-logging/0.log" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.657722 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zlfmg_bc33af41-5aa0-4254-ac75-69433d5f4ce9/ovn-controller/0.log" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.658698 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.658705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.658807 4926 scope.go:117] "RemoveContainer" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.658555 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" exitCode=0 Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660111 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" exitCode=0 Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660138 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" exitCode=0 Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660160 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" exitCode=0 Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660180 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" exitCode=0 Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660199 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" containerID="1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" exitCode=143 Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660352 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660396 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660416 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660467 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660487 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660499 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660510 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660521 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660532 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660543 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660553 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660564 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660579 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zlfmg" event={"ID":"bc33af41-5aa0-4254-ac75-69433d5f4ce9","Type":"ContainerDied","Data":"30c6b754be03f1fd88325819e9237821f35a5d4ca5f0ad8545574e86e65cadf9"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660598 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660612 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660623 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660641 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660652 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660663 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660674 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660684 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660695 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.660705 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe"} Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.685780 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.744688 4926 scope.go:117] "RemoveContainer" containerID="0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.764640 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zlfmg"] Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.769212 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zlfmg"] Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.778803 4926 scope.go:117] "RemoveContainer" containerID="c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.796647 4926 scope.go:117] "RemoveContainer" containerID="88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.813147 4926 scope.go:117] "RemoveContainer" containerID="6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.840693 4926 scope.go:117] "RemoveContainer" containerID="d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.870626 4926 scope.go:117] "RemoveContainer" containerID="1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.887265 4926 scope.go:117] "RemoveContainer" containerID="c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.920655 4926 scope.go:117] "RemoveContainer" containerID="5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.951323 4926 scope.go:117] "RemoveContainer" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.951952 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": container with ID starting with a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d not found: ID does not exist" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.952023 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} err="failed to get container status \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": rpc error: code = NotFound desc = could not find container \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": container with ID starting with a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.952067 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.952516 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": container with ID starting with 4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d not found: ID does not exist" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.952571 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} err="failed to get container status \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": rpc error: code = NotFound desc = could not find container \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": container with ID starting with 4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.952608 4926 scope.go:117] "RemoveContainer" containerID="0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.952947 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": container with ID starting with 0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4 not found: ID does not exist" containerID="0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.952984 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} err="failed to get container status \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": rpc error: code = NotFound desc = could not find container \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": container with ID starting with 0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.953008 4926 scope.go:117] "RemoveContainer" containerID="c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.953522 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": container with ID starting with c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9 not found: ID does not exist" containerID="c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.953565 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} err="failed to get container status \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": rpc error: code = NotFound desc = could not find container \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": container with ID starting with c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.953592 4926 scope.go:117] "RemoveContainer" containerID="88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.953994 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": container with ID starting with 88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5 not found: ID does not exist" containerID="88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.954044 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} err="failed to get container status \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": rpc error: code = NotFound desc = could not find container \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": container with ID starting with 88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.954078 4926 scope.go:117] "RemoveContainer" containerID="6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.954410 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": container with ID starting with 6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d not found: ID does not exist" containerID="6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.954477 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} err="failed to get container status \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": rpc error: code = NotFound desc = could not find container \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": container with ID starting with 6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.954504 4926 scope.go:117] "RemoveContainer" containerID="d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.954823 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": container with ID starting with d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a not found: ID does not exist" containerID="d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.954854 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} err="failed to get container status \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": rpc error: code = NotFound desc = could not find container \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": container with ID starting with d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.954871 4926 scope.go:117] "RemoveContainer" containerID="1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.955145 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": container with ID starting with 1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171 not found: ID does not exist" containerID="1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.955186 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} err="failed to get container status \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": rpc error: code = NotFound desc = could not find container \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": container with ID starting with 1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.955211 4926 scope.go:117] "RemoveContainer" containerID="c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.955631 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": container with ID starting with c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2 not found: ID does not exist" containerID="c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.955658 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} err="failed to get container status \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": rpc error: code = NotFound desc = could not find container \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": container with ID starting with c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.955677 4926 scope.go:117] "RemoveContainer" containerID="5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe" Mar 12 18:15:54 crc kubenswrapper[4926]: E0312 18:15:54.956039 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": container with ID starting with 5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe not found: ID does not exist" containerID="5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.956088 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe"} err="failed to get container status \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": rpc error: code = NotFound desc = could not find container \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": container with ID starting with 5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.956116 4926 scope.go:117] "RemoveContainer" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.956425 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} err="failed to get container status \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": rpc error: code = NotFound desc = could not find container \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": container with ID starting with a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.956506 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.956959 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} err="failed to get container status \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": rpc error: code = NotFound desc = could not find container \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": container with ID starting with 4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.956997 4926 scope.go:117] "RemoveContainer" containerID="0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.957546 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} err="failed to get container status \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": rpc error: code = NotFound desc = could not find container \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": container with ID starting with 0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.957574 4926 scope.go:117] "RemoveContainer" containerID="c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.957902 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} err="failed to get container status \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": rpc error: code = NotFound desc = could not find container \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": container with ID starting with c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.957943 4926 scope.go:117] "RemoveContainer" containerID="88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.958326 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} err="failed to get container status \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": rpc error: code = NotFound desc = could not find container \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": container with ID starting with 88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.958352 4926 scope.go:117] "RemoveContainer" containerID="6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.958711 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} err="failed to get container status \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": rpc error: code = NotFound desc = could not find container \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": container with ID starting with 6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.958747 4926 scope.go:117] "RemoveContainer" containerID="d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.959064 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} err="failed to get container status \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": rpc error: code = NotFound desc = could not find container \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": container with ID starting with d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.959085 4926 scope.go:117] "RemoveContainer" containerID="1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.959380 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} err="failed to get container status \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": rpc error: code = NotFound desc = could not find container \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": container with ID starting with 1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.959416 4926 scope.go:117] "RemoveContainer" containerID="c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.959970 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} err="failed to get container status \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": rpc error: code = NotFound desc = could not find container \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": container with ID starting with c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.959999 4926 scope.go:117] "RemoveContainer" containerID="5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.960340 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe"} err="failed to get container status \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": rpc error: code = NotFound desc = could not find container \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": container with ID starting with 5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.960398 4926 scope.go:117] "RemoveContainer" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.961019 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} err="failed to get container status \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": rpc error: code = NotFound desc = could not find container \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": container with ID starting with a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.961060 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.961386 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} err="failed to get container status \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": rpc error: code = NotFound desc = could not find container \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": container with ID starting with 4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.961407 4926 scope.go:117] "RemoveContainer" containerID="0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.961916 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} err="failed to get container status \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": rpc error: code = NotFound desc = could not find container \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": container with ID starting with 0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.961960 4926 scope.go:117] "RemoveContainer" containerID="c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.962492 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} err="failed to get container status \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": rpc error: code = NotFound desc = could not find container \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": container with ID starting with c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.962518 4926 scope.go:117] "RemoveContainer" containerID="88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.962953 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} err="failed to get container status \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": rpc error: code = NotFound desc = could not find container \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": container with ID starting with 88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.962995 4926 scope.go:117] "RemoveContainer" containerID="6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.963328 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} err="failed to get container status \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": rpc error: code = NotFound desc = could not find container \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": container with ID starting with 6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.963369 4926 scope.go:117] "RemoveContainer" containerID="d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.963706 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} err="failed to get container status \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": rpc error: code = NotFound desc = could not find container \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": container with ID starting with d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.963732 4926 scope.go:117] "RemoveContainer" containerID="1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.964025 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} err="failed to get container status \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": rpc error: code = NotFound desc = could not find container \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": container with ID starting with 1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.964055 4926 scope.go:117] "RemoveContainer" containerID="c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.964394 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} err="failed to get container status \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": rpc error: code = NotFound desc = could not find container \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": container with ID starting with c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.964418 4926 scope.go:117] "RemoveContainer" containerID="5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.964744 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe"} err="failed to get container status \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": rpc error: code = NotFound desc = could not find container \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": container with ID starting with 5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.964769 4926 scope.go:117] "RemoveContainer" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.965056 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} err="failed to get container status \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": rpc error: code = NotFound desc = could not find container \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": container with ID starting with a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.965082 4926 scope.go:117] "RemoveContainer" containerID="4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.965385 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d"} err="failed to get container status \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": rpc error: code = NotFound desc = could not find container \"4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d\": container with ID starting with 4cefeacdf01b18198adfc4e477f52aa244393d6d3a5bdc1c6910c26c5054b51d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.965410 4926 scope.go:117] "RemoveContainer" containerID="0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.965831 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4"} err="failed to get container status \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": rpc error: code = NotFound desc = could not find container \"0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4\": container with ID starting with 0e7256eaa446c53af7a1b21e76d692293ff3adb9f835bd6b4d3ff47074e38ae4 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.965889 4926 scope.go:117] "RemoveContainer" containerID="c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.966361 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9"} err="failed to get container status \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": rpc error: code = NotFound desc = could not find container \"c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9\": container with ID starting with c6b99cf03ba8e56bb4a9fe6cd9590db23fcf485cc69a869191ed57d04675c8d9 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.966388 4926 scope.go:117] "RemoveContainer" containerID="88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.966687 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5"} err="failed to get container status \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": rpc error: code = NotFound desc = could not find container \"88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5\": container with ID starting with 88cb63d97c58ddb777833f105b5f3ec78639266b637e061e6da8633f850da6c5 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.966728 4926 scope.go:117] "RemoveContainer" containerID="6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.967064 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d"} err="failed to get container status \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": rpc error: code = NotFound desc = could not find container \"6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d\": container with ID starting with 6265e0fe77312fcdcf7e870ecba7e55a10f1ba96ca2e0b405a377ed75013bc4d not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.967089 4926 scope.go:117] "RemoveContainer" containerID="d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.967503 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a"} err="failed to get container status \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": rpc error: code = NotFound desc = could not find container \"d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a\": container with ID starting with d9ffae7f86cae4692b30a6bfc1d646f6fdb53cd94499e5d9aa9abd106b13973a not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.967530 4926 scope.go:117] "RemoveContainer" containerID="1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.967850 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171"} err="failed to get container status \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": rpc error: code = NotFound desc = could not find container \"1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171\": container with ID starting with 1aadf309eb99a2cedba4cd2d2ceaf318e9f6c500624dd823ef281c3bbcbe0171 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.967874 4926 scope.go:117] "RemoveContainer" containerID="c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.968256 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2"} err="failed to get container status \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": rpc error: code = NotFound desc = could not find container \"c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2\": container with ID starting with c34ee146517b1408684a7c66b6ce58b98359e200020a8e36bd575e851053f8c2 not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.968283 4926 scope.go:117] "RemoveContainer" containerID="5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.968628 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe"} err="failed to get container status \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": rpc error: code = NotFound desc = could not find container \"5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe\": container with ID starting with 5722e101b3372b44b5f172d42ac38d0ddacf3abc51e97a696929af35b2600afe not found: ID does not exist" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.968668 4926 scope.go:117] "RemoveContainer" containerID="a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d" Mar 12 18:15:54 crc kubenswrapper[4926]: I0312 18:15:54.969034 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d"} err="failed to get container status \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": rpc error: code = NotFound desc = could not find container \"a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d\": container with ID starting with a6cd6dfb15f31ba734147337f04aacdd6c5480a9c23f974d37ab3b8e0d87f14d not found: ID does not exist" Mar 12 18:15:55 crc kubenswrapper[4926]: I0312 18:15:55.671256 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"e38b783fd4740c7da86f3b4c83e78bb8d8e06623f5e9efef89b65497cd4c3240"} Mar 12 18:15:55 crc kubenswrapper[4926]: I0312 18:15:55.671299 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"1a6b357d88744580e2fbeb41ce6b54b488057a60bc9995c8e9d3fb59a5afce9c"} Mar 12 18:15:55 crc kubenswrapper[4926]: I0312 18:15:55.671312 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"4beb46b32dc9802cd09149dd05a9d0029b8f0a7a397ba3a3dcd17c57630bd80d"} Mar 12 18:15:55 crc kubenswrapper[4926]: I0312 18:15:55.671321 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"cae0d92edf6eac9bf7d1bdf4b6f463b7a8036d5208b2af9b9e18546f51420e48"} Mar 12 18:15:55 crc kubenswrapper[4926]: I0312 18:15:55.671330 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"8585cd738fcdf6a8688c33a04792f8ba0d13f3fea2b2a2b94964b18cc5215665"} Mar 12 18:15:55 crc kubenswrapper[4926]: I0312 18:15:55.671338 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"ccfdccb75afb6065904425e06da5940863d90f7a64de166e8dc69927ef86a796"} Mar 12 18:15:56 crc kubenswrapper[4926]: I0312 18:15:56.502611 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc33af41-5aa0-4254-ac75-69433d5f4ce9" path="/var/lib/kubelet/pods/bc33af41-5aa0-4254-ac75-69433d5f4ce9/volumes" Mar 12 18:15:58 crc kubenswrapper[4926]: I0312 18:15:58.700243 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"281b710e00f80b4dbb02237d5d6febfa3c44d97f35dd3bbf53e415b68a8e6873"} Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.125812 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555656-zlrgw"] Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.126915 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.128776 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.129608 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.129717 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.207299 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8kp6\" (UniqueName: \"kubernetes.io/projected/b042fb81-959e-48c0-8a9e-87bafcad2fe3-kube-api-access-t8kp6\") pod \"auto-csr-approver-29555656-zlrgw\" (UID: \"b042fb81-959e-48c0-8a9e-87bafcad2fe3\") " pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.308489 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8kp6\" (UniqueName: \"kubernetes.io/projected/b042fb81-959e-48c0-8a9e-87bafcad2fe3-kube-api-access-t8kp6\") pod \"auto-csr-approver-29555656-zlrgw\" (UID: \"b042fb81-959e-48c0-8a9e-87bafcad2fe3\") " pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.328958 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8kp6\" (UniqueName: \"kubernetes.io/projected/b042fb81-959e-48c0-8a9e-87bafcad2fe3-kube-api-access-t8kp6\") pod \"auto-csr-approver-29555656-zlrgw\" (UID: \"b042fb81-959e-48c0-8a9e-87bafcad2fe3\") " pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.446640 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.464610 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(d0c1b896e388da9a50c7240b618174e003ef98ec177a69f69180c6a03b2e7c34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.464696 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(d0c1b896e388da9a50c7240b618174e003ef98ec177a69f69180c6a03b2e7c34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.464716 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(d0c1b896e388da9a50c7240b618174e003ef98ec177a69f69180c6a03b2e7c34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.464753 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29555656-zlrgw_openshift-infra(b042fb81-959e-48c0-8a9e-87bafcad2fe3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29555656-zlrgw_openshift-infra(b042fb81-959e-48c0-8a9e-87bafcad2fe3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(d0c1b896e388da9a50c7240b618174e003ef98ec177a69f69180c6a03b2e7c34): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" podUID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.713602 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555656-zlrgw"] Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.724940 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.724960 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" event={"ID":"2054c0e9-f3e6-4d68-ab8f-57466f8b3e47","Type":"ContainerStarted","Data":"f4be59f72789562e0b01dc8bccac7a56d05ba2e07d619f4b205456956452b76d"} Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.725514 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.752432 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(9113d0c0d90bdc1bca38d13974c2f38873e294bc154dcca3f50372de298592d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.752531 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(9113d0c0d90bdc1bca38d13974c2f38873e294bc154dcca3f50372de298592d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.752563 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(9113d0c0d90bdc1bca38d13974c2f38873e294bc154dcca3f50372de298592d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:00 crc kubenswrapper[4926]: E0312 18:16:00.752629 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29555656-zlrgw_openshift-infra(b042fb81-959e-48c0-8a9e-87bafcad2fe3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29555656-zlrgw_openshift-infra(b042fb81-959e-48c0-8a9e-87bafcad2fe3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(9113d0c0d90bdc1bca38d13974c2f38873e294bc154dcca3f50372de298592d7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" podUID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" Mar 12 18:16:00 crc kubenswrapper[4926]: I0312 18:16:00.759509 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" podStartSLOduration=7.759485408 podStartE2EDuration="7.759485408s" podCreationTimestamp="2026-03-12 18:15:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:16:00.759353144 +0000 UTC m=+801.127979477" watchObservedRunningTime="2026-03-12 18:16:00.759485408 +0000 UTC m=+801.128111761" Mar 12 18:16:01 crc kubenswrapper[4926]: I0312 18:16:01.731861 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:16:01 crc kubenswrapper[4926]: I0312 18:16:01.732393 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:16:01 crc kubenswrapper[4926]: I0312 18:16:01.732417 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:16:01 crc kubenswrapper[4926]: I0312 18:16:01.775349 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:16:01 crc kubenswrapper[4926]: I0312 18:16:01.781628 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:16:04 crc kubenswrapper[4926]: I0312 18:16:04.490142 4926 scope.go:117] "RemoveContainer" containerID="2c510c831017eeb7aed88601040d4790cf6d5bebce7b09f246c92ea9b81e2481" Mar 12 18:16:04 crc kubenswrapper[4926]: E0312 18:16:04.490885 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xwqvl_openshift-multus(d5a53ef4-c701-457f-9cf2-85819bf04d1a)\"" pod="openshift-multus/multus-xwqvl" podUID="d5a53ef4-c701-457f-9cf2-85819bf04d1a" Mar 12 18:16:15 crc kubenswrapper[4926]: I0312 18:16:15.488970 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:15 crc kubenswrapper[4926]: I0312 18:16:15.490352 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:15 crc kubenswrapper[4926]: E0312 18:16:15.538040 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(ff7b5c0c8b1aa68b40982f6422f49a4b9f491894d885b9649c85dff7ed33dcc6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 18:16:15 crc kubenswrapper[4926]: E0312 18:16:15.538134 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(ff7b5c0c8b1aa68b40982f6422f49a4b9f491894d885b9649c85dff7ed33dcc6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:15 crc kubenswrapper[4926]: E0312 18:16:15.538172 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(ff7b5c0c8b1aa68b40982f6422f49a4b9f491894d885b9649c85dff7ed33dcc6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:15 crc kubenswrapper[4926]: E0312 18:16:15.538238 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29555656-zlrgw_openshift-infra(b042fb81-959e-48c0-8a9e-87bafcad2fe3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29555656-zlrgw_openshift-infra(b042fb81-959e-48c0-8a9e-87bafcad2fe3)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29555656-zlrgw_openshift-infra_b042fb81-959e-48c0-8a9e-87bafcad2fe3_0(ff7b5c0c8b1aa68b40982f6422f49a4b9f491894d885b9649c85dff7ed33dcc6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" podUID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" Mar 12 18:16:17 crc kubenswrapper[4926]: I0312 18:16:17.489951 4926 scope.go:117] "RemoveContainer" containerID="2c510c831017eeb7aed88601040d4790cf6d5bebce7b09f246c92ea9b81e2481" Mar 12 18:16:17 crc kubenswrapper[4926]: I0312 18:16:17.850611 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xwqvl_d5a53ef4-c701-457f-9cf2-85819bf04d1a/kube-multus/2.log" Mar 12 18:16:17 crc kubenswrapper[4926]: I0312 18:16:17.850690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xwqvl" event={"ID":"d5a53ef4-c701-457f-9cf2-85819bf04d1a","Type":"ContainerStarted","Data":"1b5a9c1c4921fcd84b07bfff93da37ebd99a8d7476c2c239000be3cad6b4c354"} Mar 12 18:16:24 crc kubenswrapper[4926]: I0312 18:16:24.188298 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-94d7p" Mar 12 18:16:28 crc kubenswrapper[4926]: I0312 18:16:28.489418 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:28 crc kubenswrapper[4926]: I0312 18:16:28.490650 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:28 crc kubenswrapper[4926]: I0312 18:16:28.766017 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555656-zlrgw"] Mar 12 18:16:28 crc kubenswrapper[4926]: I0312 18:16:28.935391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" event={"ID":"b042fb81-959e-48c0-8a9e-87bafcad2fe3","Type":"ContainerStarted","Data":"3d3544bca4bb5b1538f3d1d62f8c4d4b5d1bdd2214c7ce446ae4aa62d0f37b88"} Mar 12 18:16:30 crc kubenswrapper[4926]: I0312 18:16:30.959289 4926 generic.go:334] "Generic (PLEG): container finished" podID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" containerID="d82acd09ec6775228cff799be28fec22c80e7a284a7159b9a2b0821199a8fa17" exitCode=0 Mar 12 18:16:30 crc kubenswrapper[4926]: I0312 18:16:30.959404 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" event={"ID":"b042fb81-959e-48c0-8a9e-87bafcad2fe3","Type":"ContainerDied","Data":"d82acd09ec6775228cff799be28fec22c80e7a284a7159b9a2b0821199a8fa17"} Mar 12 18:16:32 crc kubenswrapper[4926]: I0312 18:16:32.246247 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:32 crc kubenswrapper[4926]: I0312 18:16:32.357005 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8kp6\" (UniqueName: \"kubernetes.io/projected/b042fb81-959e-48c0-8a9e-87bafcad2fe3-kube-api-access-t8kp6\") pod \"b042fb81-959e-48c0-8a9e-87bafcad2fe3\" (UID: \"b042fb81-959e-48c0-8a9e-87bafcad2fe3\") " Mar 12 18:16:32 crc kubenswrapper[4926]: I0312 18:16:32.370695 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b042fb81-959e-48c0-8a9e-87bafcad2fe3-kube-api-access-t8kp6" (OuterVolumeSpecName: "kube-api-access-t8kp6") pod "b042fb81-959e-48c0-8a9e-87bafcad2fe3" (UID: "b042fb81-959e-48c0-8a9e-87bafcad2fe3"). InnerVolumeSpecName "kube-api-access-t8kp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:16:32 crc kubenswrapper[4926]: I0312 18:16:32.458511 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8kp6\" (UniqueName: \"kubernetes.io/projected/b042fb81-959e-48c0-8a9e-87bafcad2fe3-kube-api-access-t8kp6\") on node \"crc\" DevicePath \"\"" Mar 12 18:16:32 crc kubenswrapper[4926]: I0312 18:16:32.969646 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" event={"ID":"b042fb81-959e-48c0-8a9e-87bafcad2fe3","Type":"ContainerDied","Data":"3d3544bca4bb5b1538f3d1d62f8c4d4b5d1bdd2214c7ce446ae4aa62d0f37b88"} Mar 12 18:16:32 crc kubenswrapper[4926]: I0312 18:16:32.969902 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d3544bca4bb5b1538f3d1d62f8c4d4b5d1bdd2214c7ce446ae4aa62d0f37b88" Mar 12 18:16:32 crc kubenswrapper[4926]: I0312 18:16:32.969692 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555656-zlrgw" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.210115 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq"] Mar 12 18:16:33 crc kubenswrapper[4926]: E0312 18:16:33.210315 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" containerName="oc" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.210327 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" containerName="oc" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.210456 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" containerName="oc" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.211212 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.213065 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.222524 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq"] Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.291757 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555650-t22sq"] Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.295229 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555650-t22sq"] Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.375629 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.376197 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.376308 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97jtx\" (UniqueName: \"kubernetes.io/projected/062d1c31-cb0c-4470-bf74-0fb541319609-kube-api-access-97jtx\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.477900 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.478026 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.478072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97jtx\" (UniqueName: \"kubernetes.io/projected/062d1c31-cb0c-4470-bf74-0fb541319609-kube-api-access-97jtx\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.478590 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.478946 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.503615 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97jtx\" (UniqueName: \"kubernetes.io/projected/062d1c31-cb0c-4470-bf74-0fb541319609-kube-api-access-97jtx\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:33 crc kubenswrapper[4926]: I0312 18:16:33.547144 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:34 crc kubenswrapper[4926]: I0312 18:16:34.008810 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq"] Mar 12 18:16:34 crc kubenswrapper[4926]: I0312 18:16:34.502935 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a5ff64f-2478-4592-8a08-fb47a40a8de5" path="/var/lib/kubelet/pods/5a5ff64f-2478-4592-8a08-fb47a40a8de5/volumes" Mar 12 18:16:34 crc kubenswrapper[4926]: I0312 18:16:34.985991 4926 generic.go:334] "Generic (PLEG): container finished" podID="062d1c31-cb0c-4470-bf74-0fb541319609" containerID="fc8115a8d2e23476b7d9888864d7a849fc584d14e0bc6160ffd2c7b9d5daad64" exitCode=0 Mar 12 18:16:34 crc kubenswrapper[4926]: I0312 18:16:34.986064 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" event={"ID":"062d1c31-cb0c-4470-bf74-0fb541319609","Type":"ContainerDied","Data":"fc8115a8d2e23476b7d9888864d7a849fc584d14e0bc6160ffd2c7b9d5daad64"} Mar 12 18:16:34 crc kubenswrapper[4926]: I0312 18:16:34.986111 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" event={"ID":"062d1c31-cb0c-4470-bf74-0fb541319609","Type":"ContainerStarted","Data":"3b98f38c5476d8f4dc24270676c0499d10c24a9bd4a1b327a146aba83d08aa26"} Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.311571 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h5l25"] Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.313614 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.324231 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h5l25"] Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.505522 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-575gn\" (UniqueName: \"kubernetes.io/projected/1ec7f260-07a4-4602-bba9-c08f20f8cfde-kube-api-access-575gn\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.505582 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-catalog-content\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.505618 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-utilities\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.606796 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-utilities\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.606967 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-575gn\" (UniqueName: \"kubernetes.io/projected/1ec7f260-07a4-4602-bba9-c08f20f8cfde-kube-api-access-575gn\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.607028 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-catalog-content\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.607424 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-utilities\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.607599 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-catalog-content\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.626412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-575gn\" (UniqueName: \"kubernetes.io/projected/1ec7f260-07a4-4602-bba9-c08f20f8cfde-kube-api-access-575gn\") pod \"redhat-operators-h5l25\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.639193 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.839280 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h5l25"] Mar 12 18:16:35 crc kubenswrapper[4926]: W0312 18:16:35.846803 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ec7f260_07a4_4602_bba9_c08f20f8cfde.slice/crio-2a487c7409c3ac193f99d036af52bf58251bf6af395bf230d56d515dbbd21905 WatchSource:0}: Error finding container 2a487c7409c3ac193f99d036af52bf58251bf6af395bf230d56d515dbbd21905: Status 404 returned error can't find the container with id 2a487c7409c3ac193f99d036af52bf58251bf6af395bf230d56d515dbbd21905 Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.992372 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerStarted","Data":"77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6"} Mar 12 18:16:35 crc kubenswrapper[4926]: I0312 18:16:35.992410 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerStarted","Data":"2a487c7409c3ac193f99d036af52bf58251bf6af395bf230d56d515dbbd21905"} Mar 12 18:16:37 crc kubenswrapper[4926]: I0312 18:16:37.007853 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerID="77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6" exitCode=0 Mar 12 18:16:37 crc kubenswrapper[4926]: I0312 18:16:37.007980 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerDied","Data":"77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6"} Mar 12 18:16:37 crc kubenswrapper[4926]: I0312 18:16:37.810545 4926 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 12 18:16:38 crc kubenswrapper[4926]: I0312 18:16:38.017797 4926 generic.go:334] "Generic (PLEG): container finished" podID="062d1c31-cb0c-4470-bf74-0fb541319609" containerID="478e8bb351f2aa6703215d63de484eb3a3adde69425a8c06e70fa114ba1298de" exitCode=0 Mar 12 18:16:38 crc kubenswrapper[4926]: I0312 18:16:38.017903 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" event={"ID":"062d1c31-cb0c-4470-bf74-0fb541319609","Type":"ContainerDied","Data":"478e8bb351f2aa6703215d63de484eb3a3adde69425a8c06e70fa114ba1298de"} Mar 12 18:16:38 crc kubenswrapper[4926]: I0312 18:16:38.020629 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerStarted","Data":"74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320"} Mar 12 18:16:39 crc kubenswrapper[4926]: I0312 18:16:39.036725 4926 generic.go:334] "Generic (PLEG): container finished" podID="062d1c31-cb0c-4470-bf74-0fb541319609" containerID="02f4293db8ec40c24e25998ced5750ff8c11823ccd164966d00cb2b50fd60c19" exitCode=0 Mar 12 18:16:39 crc kubenswrapper[4926]: I0312 18:16:39.037252 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" event={"ID":"062d1c31-cb0c-4470-bf74-0fb541319609","Type":"ContainerDied","Data":"02f4293db8ec40c24e25998ced5750ff8c11823ccd164966d00cb2b50fd60c19"} Mar 12 18:16:39 crc kubenswrapper[4926]: I0312 18:16:39.047393 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerID="74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320" exitCode=0 Mar 12 18:16:39 crc kubenswrapper[4926]: I0312 18:16:39.047475 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerDied","Data":"74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320"} Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.058601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerStarted","Data":"cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5"} Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.087839 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h5l25" podStartSLOduration=2.556300414 podStartE2EDuration="5.087816594s" podCreationTimestamp="2026-03-12 18:16:35 +0000 UTC" firstStartedPulling="2026-03-12 18:16:37.010360518 +0000 UTC m=+837.378986881" lastFinishedPulling="2026-03-12 18:16:39.541876718 +0000 UTC m=+839.910503061" observedRunningTime="2026-03-12 18:16:40.085050618 +0000 UTC m=+840.453676951" watchObservedRunningTime="2026-03-12 18:16:40.087816594 +0000 UTC m=+840.456442977" Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.405713 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.482811 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-bundle\") pod \"062d1c31-cb0c-4470-bf74-0fb541319609\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.482874 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97jtx\" (UniqueName: \"kubernetes.io/projected/062d1c31-cb0c-4470-bf74-0fb541319609-kube-api-access-97jtx\") pod \"062d1c31-cb0c-4470-bf74-0fb541319609\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.483044 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-util\") pod \"062d1c31-cb0c-4470-bf74-0fb541319609\" (UID: \"062d1c31-cb0c-4470-bf74-0fb541319609\") " Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.484018 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-bundle" (OuterVolumeSpecName: "bundle") pod "062d1c31-cb0c-4470-bf74-0fb541319609" (UID: "062d1c31-cb0c-4470-bf74-0fb541319609"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.490609 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/062d1c31-cb0c-4470-bf74-0fb541319609-kube-api-access-97jtx" (OuterVolumeSpecName: "kube-api-access-97jtx") pod "062d1c31-cb0c-4470-bf74-0fb541319609" (UID: "062d1c31-cb0c-4470-bf74-0fb541319609"). InnerVolumeSpecName "kube-api-access-97jtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.496905 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-util" (OuterVolumeSpecName: "util") pod "062d1c31-cb0c-4470-bf74-0fb541319609" (UID: "062d1c31-cb0c-4470-bf74-0fb541319609"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.584249 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-util\") on node \"crc\" DevicePath \"\"" Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.584295 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/062d1c31-cb0c-4470-bf74-0fb541319609-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:16:40 crc kubenswrapper[4926]: I0312 18:16:40.584314 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97jtx\" (UniqueName: \"kubernetes.io/projected/062d1c31-cb0c-4470-bf74-0fb541319609-kube-api-access-97jtx\") on node \"crc\" DevicePath \"\"" Mar 12 18:16:41 crc kubenswrapper[4926]: I0312 18:16:41.069080 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" Mar 12 18:16:41 crc kubenswrapper[4926]: I0312 18:16:41.069267 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq" event={"ID":"062d1c31-cb0c-4470-bf74-0fb541319609","Type":"ContainerDied","Data":"3b98f38c5476d8f4dc24270676c0499d10c24a9bd4a1b327a146aba83d08aa26"} Mar 12 18:16:41 crc kubenswrapper[4926]: I0312 18:16:41.069808 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b98f38c5476d8f4dc24270676c0499d10c24a9bd4a1b327a146aba83d08aa26" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.601743 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv"] Mar 12 18:16:43 crc kubenswrapper[4926]: E0312 18:16:43.601992 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062d1c31-cb0c-4470-bf74-0fb541319609" containerName="pull" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.602006 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="062d1c31-cb0c-4470-bf74-0fb541319609" containerName="pull" Mar 12 18:16:43 crc kubenswrapper[4926]: E0312 18:16:43.602028 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062d1c31-cb0c-4470-bf74-0fb541319609" containerName="util" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.602036 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="062d1c31-cb0c-4470-bf74-0fb541319609" containerName="util" Mar 12 18:16:43 crc kubenswrapper[4926]: E0312 18:16:43.602049 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062d1c31-cb0c-4470-bf74-0fb541319609" containerName="extract" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.602059 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="062d1c31-cb0c-4470-bf74-0fb541319609" containerName="extract" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.602184 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="062d1c31-cb0c-4470-bf74-0fb541319609" containerName="extract" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.602658 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.604400 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.604655 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.610010 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-695zc" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.611719 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv"] Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.624930 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c4gt\" (UniqueName: \"kubernetes.io/projected/7247f6a4-e62d-44e3-b91c-1117fca5c960-kube-api-access-9c4gt\") pod \"nmstate-operator-796d4cfff4-wnkdv\" (UID: \"7247f6a4-e62d-44e3-b91c-1117fca5c960\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.735380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c4gt\" (UniqueName: \"kubernetes.io/projected/7247f6a4-e62d-44e3-b91c-1117fca5c960-kube-api-access-9c4gt\") pod \"nmstate-operator-796d4cfff4-wnkdv\" (UID: \"7247f6a4-e62d-44e3-b91c-1117fca5c960\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.753526 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c4gt\" (UniqueName: \"kubernetes.io/projected/7247f6a4-e62d-44e3-b91c-1117fca5c960-kube-api-access-9c4gt\") pod \"nmstate-operator-796d4cfff4-wnkdv\" (UID: \"7247f6a4-e62d-44e3-b91c-1117fca5c960\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" Mar 12 18:16:43 crc kubenswrapper[4926]: I0312 18:16:43.916989 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" Mar 12 18:16:44 crc kubenswrapper[4926]: I0312 18:16:44.169277 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv"] Mar 12 18:16:44 crc kubenswrapper[4926]: W0312 18:16:44.172495 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7247f6a4_e62d_44e3_b91c_1117fca5c960.slice/crio-5799508215f1be1b36b37688c4128bd4007438ec6aab48f2b9928063c5f1a50f WatchSource:0}: Error finding container 5799508215f1be1b36b37688c4128bd4007438ec6aab48f2b9928063c5f1a50f: Status 404 returned error can't find the container with id 5799508215f1be1b36b37688c4128bd4007438ec6aab48f2b9928063c5f1a50f Mar 12 18:16:45 crc kubenswrapper[4926]: I0312 18:16:45.098284 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" event={"ID":"7247f6a4-e62d-44e3-b91c-1117fca5c960","Type":"ContainerStarted","Data":"5799508215f1be1b36b37688c4128bd4007438ec6aab48f2b9928063c5f1a50f"} Mar 12 18:16:45 crc kubenswrapper[4926]: I0312 18:16:45.639883 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:45 crc kubenswrapper[4926]: I0312 18:16:45.640279 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:46 crc kubenswrapper[4926]: I0312 18:16:46.682863 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h5l25" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="registry-server" probeResult="failure" output=< Mar 12 18:16:46 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:16:46 crc kubenswrapper[4926]: > Mar 12 18:16:47 crc kubenswrapper[4926]: I0312 18:16:47.115316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" event={"ID":"7247f6a4-e62d-44e3-b91c-1117fca5c960","Type":"ContainerStarted","Data":"c53d52de6035183f9ebe6f6214ef9eec047645edbfcc2112636630f37a358302"} Mar 12 18:16:47 crc kubenswrapper[4926]: I0312 18:16:47.144670 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-wnkdv" podStartSLOduration=1.6531536610000002 podStartE2EDuration="4.144652096s" podCreationTimestamp="2026-03-12 18:16:43 +0000 UTC" firstStartedPulling="2026-03-12 18:16:44.174792058 +0000 UTC m=+844.543418411" lastFinishedPulling="2026-03-12 18:16:46.666290513 +0000 UTC m=+847.034916846" observedRunningTime="2026-03-12 18:16:47.141816728 +0000 UTC m=+847.510443071" watchObservedRunningTime="2026-03-12 18:16:47.144652096 +0000 UTC m=+847.513278429" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.197479 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.199479 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.202164 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-9nqhs" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.213874 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.244275 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.247597 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.256312 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.271680 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-z52rq"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.272686 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.273553 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkbcg\" (UniqueName: \"kubernetes.io/projected/cf97acff-2650-4e84-ab30-d10f9bd70ef4-kube-api-access-dkbcg\") pod \"nmstate-metrics-9b8c8685d-rn7dj\" (UID: \"cf97acff-2650-4e84-ab30-d10f9bd70ef4\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.273614 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-h5ll8\" (UID: \"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.273729 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kllkp\" (UniqueName: \"kubernetes.io/projected/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-kube-api-access-kllkp\") pod \"nmstate-webhook-5f558f5558-h5ll8\" (UID: \"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.301690 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.342691 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.343564 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.346829 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.347029 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-ptr59" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.347208 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.350745 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375306 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375360 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-dbus-socket\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375386 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-nmstate-lock\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375466 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-ovs-socket\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375515 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kllkp\" (UniqueName: \"kubernetes.io/projected/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-kube-api-access-kllkp\") pod \"nmstate-webhook-5f558f5558-h5ll8\" (UID: \"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375554 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375592 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkbcg\" (UniqueName: \"kubernetes.io/projected/cf97acff-2650-4e84-ab30-d10f9bd70ef4-kube-api-access-dkbcg\") pod \"nmstate-metrics-9b8c8685d-rn7dj\" (UID: \"cf97acff-2650-4e84-ab30-d10f9bd70ef4\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375615 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg5fc\" (UniqueName: \"kubernetes.io/projected/9db224ca-f640-4756-9c80-afe7ff63dcbe-kube-api-access-hg5fc\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375634 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-h5ll8\" (UID: \"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.375650 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2spkt\" (UniqueName: \"kubernetes.io/projected/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-kube-api-access-2spkt\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: E0312 18:16:53.375830 4926 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 12 18:16:53 crc kubenswrapper[4926]: E0312 18:16:53.375878 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-tls-key-pair podName:bdc3504f-bf1b-4b71-aaf2-45e24e41a84e nodeName:}" failed. No retries permitted until 2026-03-12 18:16:53.875862522 +0000 UTC m=+854.244488855 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-tls-key-pair") pod "nmstate-webhook-5f558f5558-h5ll8" (UID: "bdc3504f-bf1b-4b71-aaf2-45e24e41a84e") : secret "openshift-nmstate-webhook" not found Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.394767 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kllkp\" (UniqueName: \"kubernetes.io/projected/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-kube-api-access-kllkp\") pod \"nmstate-webhook-5f558f5558-h5ll8\" (UID: \"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.396572 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkbcg\" (UniqueName: \"kubernetes.io/projected/cf97acff-2650-4e84-ab30-d10f9bd70ef4-kube-api-access-dkbcg\") pod \"nmstate-metrics-9b8c8685d-rn7dj\" (UID: \"cf97acff-2650-4e84-ab30-d10f9bd70ef4\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477593 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-dbus-socket\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: E0312 18:16:53.477301 4926 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477643 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-nmstate-lock\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477671 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-ovs-socket\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: E0312 18:16:53.477697 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-plugin-serving-cert podName:9e731d72-a0a8-46a4-af0b-5dce65f29dd1 nodeName:}" failed. No retries permitted until 2026-03-12 18:16:53.97767439 +0000 UTC m=+854.346300733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-67hvb" (UID: "9e731d72-a0a8-46a4-af0b-5dce65f29dd1") : secret "plugin-serving-cert" not found Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477714 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477787 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg5fc\" (UniqueName: \"kubernetes.io/projected/9db224ca-f640-4756-9c80-afe7ff63dcbe-kube-api-access-hg5fc\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2spkt\" (UniqueName: \"kubernetes.io/projected/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-kube-api-access-2spkt\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.477978 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-ovs-socket\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.478162 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-nmstate-lock\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.478266 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/9db224ca-f640-4756-9c80-afe7ff63dcbe-dbus-socket\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.479047 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.497102 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg5fc\" (UniqueName: \"kubernetes.io/projected/9db224ca-f640-4756-9c80-afe7ff63dcbe-kube-api-access-hg5fc\") pod \"nmstate-handler-z52rq\" (UID: \"9db224ca-f640-4756-9c80-afe7ff63dcbe\") " pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.497177 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2spkt\" (UniqueName: \"kubernetes.io/projected/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-kube-api-access-2spkt\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.518235 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.525204 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5dd74c7c8d-h748s"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.526063 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.577507 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dd74c7c8d-h748s"] Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.578780 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-trusted-ca-bundle\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.578841 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-oauth-serving-cert\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.578870 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-console-config\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.578892 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-service-ca\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.578930 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2872acc-add0-4247-b9de-29182d576add-console-serving-cert\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.578980 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99fkz\" (UniqueName: \"kubernetes.io/projected/a2872acc-add0-4247-b9de-29182d576add-kube-api-access-99fkz\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.579048 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2872acc-add0-4247-b9de-29182d576add-console-oauth-config\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.602049 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:53 crc kubenswrapper[4926]: W0312 18:16:53.621687 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db224ca_f640_4756_9c80_afe7ff63dcbe.slice/crio-af86013ef6d8eabf4ac321f53ed9fbc379671a4d43fa8c4ddfa2f271033910d1 WatchSource:0}: Error finding container af86013ef6d8eabf4ac321f53ed9fbc379671a4d43fa8c4ddfa2f271033910d1: Status 404 returned error can't find the container with id af86013ef6d8eabf4ac321f53ed9fbc379671a4d43fa8c4ddfa2f271033910d1 Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.680433 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-oauth-serving-cert\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.680733 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-console-config\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.680756 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-service-ca\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.680780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2872acc-add0-4247-b9de-29182d576add-console-serving-cert\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.680825 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99fkz\" (UniqueName: \"kubernetes.io/projected/a2872acc-add0-4247-b9de-29182d576add-kube-api-access-99fkz\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.680876 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2872acc-add0-4247-b9de-29182d576add-console-oauth-config\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.680933 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-trusted-ca-bundle\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.682320 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-trusted-ca-bundle\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.682576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-oauth-serving-cert\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.683173 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-console-config\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.683768 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2872acc-add0-4247-b9de-29182d576add-service-ca\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.687168 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2872acc-add0-4247-b9de-29182d576add-console-oauth-config\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.687546 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2872acc-add0-4247-b9de-29182d576add-console-serving-cert\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.714230 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99fkz\" (UniqueName: \"kubernetes.io/projected/a2872acc-add0-4247-b9de-29182d576add-kube-api-access-99fkz\") pod \"console-5dd74c7c8d-h748s\" (UID: \"a2872acc-add0-4247-b9de-29182d576add\") " pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.867208 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.882645 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-h5ll8\" (UID: \"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.886833 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bdc3504f-bf1b-4b71-aaf2-45e24e41a84e-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-h5ll8\" (UID: \"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.985047 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:53 crc kubenswrapper[4926]: I0312 18:16:53.990993 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9e731d72-a0a8-46a4-af0b-5dce65f29dd1-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-67hvb\" (UID: \"9e731d72-a0a8-46a4-af0b-5dce65f29dd1\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.018956 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj"] Mar 12 18:16:54 crc kubenswrapper[4926]: W0312 18:16:54.085948 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2872acc_add0_4247_b9de_29182d576add.slice/crio-0dd02ac57b1182a64b9e652af6125f6d4b80a1fa0283d2cf0121b614df75be64 WatchSource:0}: Error finding container 0dd02ac57b1182a64b9e652af6125f6d4b80a1fa0283d2cf0121b614df75be64: Status 404 returned error can't find the container with id 0dd02ac57b1182a64b9e652af6125f6d4b80a1fa0283d2cf0121b614df75be64 Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.085967 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dd74c7c8d-h748s"] Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.162267 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dd74c7c8d-h748s" event={"ID":"a2872acc-add0-4247-b9de-29182d576add","Type":"ContainerStarted","Data":"0dd02ac57b1182a64b9e652af6125f6d4b80a1fa0283d2cf0121b614df75be64"} Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.163720 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-z52rq" event={"ID":"9db224ca-f640-4756-9c80-afe7ff63dcbe","Type":"ContainerStarted","Data":"af86013ef6d8eabf4ac321f53ed9fbc379671a4d43fa8c4ddfa2f271033910d1"} Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.165328 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" event={"ID":"cf97acff-2650-4e84-ab30-d10f9bd70ef4","Type":"ContainerStarted","Data":"d4fa75e5be6ba0fe4bfed768e68798f6fff5cf6e00898b5ea44fe66e500455d4"} Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.176911 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.258073 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.474180 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb"] Mar 12 18:16:54 crc kubenswrapper[4926]: W0312 18:16:54.480582 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e731d72_a0a8_46a4_af0b_5dce65f29dd1.slice/crio-f147a0ab76c9688735ad96ef60ec96bc8652f2bfbfd1da41905f0fa1163a475f WatchSource:0}: Error finding container f147a0ab76c9688735ad96ef60ec96bc8652f2bfbfd1da41905f0fa1163a475f: Status 404 returned error can't find the container with id f147a0ab76c9688735ad96ef60ec96bc8652f2bfbfd1da41905f0fa1163a475f Mar 12 18:16:54 crc kubenswrapper[4926]: I0312 18:16:54.594343 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8"] Mar 12 18:16:54 crc kubenswrapper[4926]: W0312 18:16:54.603760 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdc3504f_bf1b_4b71_aaf2_45e24e41a84e.slice/crio-10294b79b92d5b6599d2e7bdd3189cfd8c9071c3e5890eaf69101a049a18c76d WatchSource:0}: Error finding container 10294b79b92d5b6599d2e7bdd3189cfd8c9071c3e5890eaf69101a049a18c76d: Status 404 returned error can't find the container with id 10294b79b92d5b6599d2e7bdd3189cfd8c9071c3e5890eaf69101a049a18c76d Mar 12 18:16:55 crc kubenswrapper[4926]: I0312 18:16:55.174054 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dd74c7c8d-h748s" event={"ID":"a2872acc-add0-4247-b9de-29182d576add","Type":"ContainerStarted","Data":"409ad4780a657ca30f7de6f8e1fd1934f3400a8d18525ae3304ebe2a7c68ca46"} Mar 12 18:16:55 crc kubenswrapper[4926]: I0312 18:16:55.175178 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" event={"ID":"9e731d72-a0a8-46a4-af0b-5dce65f29dd1","Type":"ContainerStarted","Data":"f147a0ab76c9688735ad96ef60ec96bc8652f2bfbfd1da41905f0fa1163a475f"} Mar 12 18:16:55 crc kubenswrapper[4926]: I0312 18:16:55.177113 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" event={"ID":"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e","Type":"ContainerStarted","Data":"10294b79b92d5b6599d2e7bdd3189cfd8c9071c3e5890eaf69101a049a18c76d"} Mar 12 18:16:55 crc kubenswrapper[4926]: I0312 18:16:55.198131 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5dd74c7c8d-h748s" podStartSLOduration=2.198113346 podStartE2EDuration="2.198113346s" podCreationTimestamp="2026-03-12 18:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:16:55.193484002 +0000 UTC m=+855.562110335" watchObservedRunningTime="2026-03-12 18:16:55.198113346 +0000 UTC m=+855.566739689" Mar 12 18:16:55 crc kubenswrapper[4926]: I0312 18:16:55.692456 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:55 crc kubenswrapper[4926]: I0312 18:16:55.740405 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:55 crc kubenswrapper[4926]: I0312 18:16:55.934588 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h5l25"] Mar 12 18:16:56 crc kubenswrapper[4926]: I0312 18:16:56.817306 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:16:56 crc kubenswrapper[4926]: I0312 18:16:56.817364 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.190976 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" event={"ID":"bdc3504f-bf1b-4b71-aaf2-45e24e41a84e","Type":"ContainerStarted","Data":"f9ad9d26f11464ab92e6a67ee91b0719f50bb451d1c1460ba378a94a95476530"} Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.191097 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.192580 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-z52rq" event={"ID":"9db224ca-f640-4756-9c80-afe7ff63dcbe","Type":"ContainerStarted","Data":"436dde831c9f9d785be41544e9dacc6364c03710a7397c86cc5b9441ff5b331f"} Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.192831 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.194481 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" event={"ID":"cf97acff-2650-4e84-ab30-d10f9bd70ef4","Type":"ContainerStarted","Data":"ec1620f8d6850b5716c49100add4fec7a885301893db061d8983844b50be4a57"} Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.194625 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h5l25" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="registry-server" containerID="cri-o://cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5" gracePeriod=2 Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.213361 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" podStartSLOduration=2.521425055 podStartE2EDuration="4.213340885s" podCreationTimestamp="2026-03-12 18:16:53 +0000 UTC" firstStartedPulling="2026-03-12 18:16:54.605204869 +0000 UTC m=+854.973831202" lastFinishedPulling="2026-03-12 18:16:56.297120699 +0000 UTC m=+856.665747032" observedRunningTime="2026-03-12 18:16:57.205884302 +0000 UTC m=+857.574510635" watchObservedRunningTime="2026-03-12 18:16:57.213340885 +0000 UTC m=+857.581967218" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.227252 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-z52rq" podStartSLOduration=1.554460382 podStartE2EDuration="4.227235767s" podCreationTimestamp="2026-03-12 18:16:53 +0000 UTC" firstStartedPulling="2026-03-12 18:16:53.623369683 +0000 UTC m=+853.991996026" lastFinishedPulling="2026-03-12 18:16:56.296145068 +0000 UTC m=+856.664771411" observedRunningTime="2026-03-12 18:16:57.225613856 +0000 UTC m=+857.594240189" watchObservedRunningTime="2026-03-12 18:16:57.227235767 +0000 UTC m=+857.595862100" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.525990 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.535746 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-catalog-content\") pod \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.535936 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-575gn\" (UniqueName: \"kubernetes.io/projected/1ec7f260-07a4-4602-bba9-c08f20f8cfde-kube-api-access-575gn\") pod \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.536039 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-utilities\") pod \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\" (UID: \"1ec7f260-07a4-4602-bba9-c08f20f8cfde\") " Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.540109 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-utilities" (OuterVolumeSpecName: "utilities") pod "1ec7f260-07a4-4602-bba9-c08f20f8cfde" (UID: "1ec7f260-07a4-4602-bba9-c08f20f8cfde"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.547981 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec7f260-07a4-4602-bba9-c08f20f8cfde-kube-api-access-575gn" (OuterVolumeSpecName: "kube-api-access-575gn") pod "1ec7f260-07a4-4602-bba9-c08f20f8cfde" (UID: "1ec7f260-07a4-4602-bba9-c08f20f8cfde"). InnerVolumeSpecName "kube-api-access-575gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.638025 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-575gn\" (UniqueName: \"kubernetes.io/projected/1ec7f260-07a4-4602-bba9-c08f20f8cfde-kube-api-access-575gn\") on node \"crc\" DevicePath \"\"" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.638052 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.692228 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ec7f260-07a4-4602-bba9-c08f20f8cfde" (UID: "1ec7f260-07a4-4602-bba9-c08f20f8cfde"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:16:57 crc kubenswrapper[4926]: I0312 18:16:57.739526 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ec7f260-07a4-4602-bba9-c08f20f8cfde-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.208088 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerID="cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5" exitCode=0 Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.208237 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerDied","Data":"cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5"} Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.208280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h5l25" event={"ID":"1ec7f260-07a4-4602-bba9-c08f20f8cfde","Type":"ContainerDied","Data":"2a487c7409c3ac193f99d036af52bf58251bf6af395bf230d56d515dbbd21905"} Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.208309 4926 scope.go:117] "RemoveContainer" containerID="cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.208533 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h5l25" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.213307 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" event={"ID":"9e731d72-a0a8-46a4-af0b-5dce65f29dd1","Type":"ContainerStarted","Data":"993d2423c1bbc3b0f46463a6f134db5ee4e6d76062f2c251b25aaf8c7391ad8f"} Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.234899 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-67hvb" podStartSLOduration=2.495741106 podStartE2EDuration="5.234874076s" podCreationTimestamp="2026-03-12 18:16:53 +0000 UTC" firstStartedPulling="2026-03-12 18:16:54.482012537 +0000 UTC m=+854.850638890" lastFinishedPulling="2026-03-12 18:16:57.221145527 +0000 UTC m=+857.589771860" observedRunningTime="2026-03-12 18:16:58.227060373 +0000 UTC m=+858.595686716" watchObservedRunningTime="2026-03-12 18:16:58.234874076 +0000 UTC m=+858.603500419" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.238682 4926 scope.go:117] "RemoveContainer" containerID="74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.264619 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h5l25"] Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.268133 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h5l25"] Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.271562 4926 scope.go:117] "RemoveContainer" containerID="77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.295164 4926 scope.go:117] "RemoveContainer" containerID="cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5" Mar 12 18:16:58 crc kubenswrapper[4926]: E0312 18:16:58.295734 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5\": container with ID starting with cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5 not found: ID does not exist" containerID="cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.295766 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5"} err="failed to get container status \"cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5\": rpc error: code = NotFound desc = could not find container \"cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5\": container with ID starting with cde59c82a4f343b3652464d85c7c4f6f5de5b3c10538053ac4d1fbf861b525c5 not found: ID does not exist" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.295793 4926 scope.go:117] "RemoveContainer" containerID="74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320" Mar 12 18:16:58 crc kubenswrapper[4926]: E0312 18:16:58.296228 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320\": container with ID starting with 74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320 not found: ID does not exist" containerID="74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.296272 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320"} err="failed to get container status \"74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320\": rpc error: code = NotFound desc = could not find container \"74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320\": container with ID starting with 74c5cf37effbecfaa9e3318f6a84edaeefe25fa2f0dec8231b3281e40bc9f320 not found: ID does not exist" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.296302 4926 scope.go:117] "RemoveContainer" containerID="77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6" Mar 12 18:16:58 crc kubenswrapper[4926]: E0312 18:16:58.296705 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6\": container with ID starting with 77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6 not found: ID does not exist" containerID="77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.296734 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6"} err="failed to get container status \"77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6\": rpc error: code = NotFound desc = could not find container \"77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6\": container with ID starting with 77185795828a5381b76fc3f332b62ff134a55c6dbcf3c11bcc7dd2885f8b72b6 not found: ID does not exist" Mar 12 18:16:58 crc kubenswrapper[4926]: I0312 18:16:58.498820 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" path="/var/lib/kubelet/pods/1ec7f260-07a4-4602-bba9-c08f20f8cfde/volumes" Mar 12 18:16:59 crc kubenswrapper[4926]: I0312 18:16:59.224582 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" event={"ID":"cf97acff-2650-4e84-ab30-d10f9bd70ef4","Type":"ContainerStarted","Data":"2059feca55eaa3385483d3f33e15ac1922628218eb464dd31a38e38a262e1cfa"} Mar 12 18:17:01 crc kubenswrapper[4926]: I0312 18:17:01.125302 4926 scope.go:117] "RemoveContainer" containerID="1d8805e76417daef23d168ee9dafcbdf012b728e6e9278f288005edef5f23d20" Mar 12 18:17:03 crc kubenswrapper[4926]: I0312 18:17:03.631215 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-z52rq" Mar 12 18:17:03 crc kubenswrapper[4926]: I0312 18:17:03.653529 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-rn7dj" podStartSLOduration=5.641639991 podStartE2EDuration="10.653510991s" podCreationTimestamp="2026-03-12 18:16:53 +0000 UTC" firstStartedPulling="2026-03-12 18:16:54.018646511 +0000 UTC m=+854.387272884" lastFinishedPulling="2026-03-12 18:16:59.030517541 +0000 UTC m=+859.399143884" observedRunningTime="2026-03-12 18:16:59.247090858 +0000 UTC m=+859.615717241" watchObservedRunningTime="2026-03-12 18:17:03.653510991 +0000 UTC m=+864.022137334" Mar 12 18:17:03 crc kubenswrapper[4926]: I0312 18:17:03.867684 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:17:03 crc kubenswrapper[4926]: I0312 18:17:03.867782 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:17:03 crc kubenswrapper[4926]: I0312 18:17:03.876056 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:17:04 crc kubenswrapper[4926]: I0312 18:17:04.270684 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5dd74c7c8d-h748s" Mar 12 18:17:04 crc kubenswrapper[4926]: I0312 18:17:04.323256 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vb9qx"] Mar 12 18:17:14 crc kubenswrapper[4926]: I0312 18:17:14.186512 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-h5ll8" Mar 12 18:17:26 crc kubenswrapper[4926]: I0312 18:17:26.817693 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:17:26 crc kubenswrapper[4926]: I0312 18:17:26.818643 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.194739 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg"] Mar 12 18:17:29 crc kubenswrapper[4926]: E0312 18:17:29.197968 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="extract-utilities" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.197994 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="extract-utilities" Mar 12 18:17:29 crc kubenswrapper[4926]: E0312 18:17:29.198005 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="extract-content" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.198014 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="extract-content" Mar 12 18:17:29 crc kubenswrapper[4926]: E0312 18:17:29.198045 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="registry-server" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.198055 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="registry-server" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.198201 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec7f260-07a4-4602-bba9-c08f20f8cfde" containerName="registry-server" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.199204 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.199355 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg"] Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.200902 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.278897 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.278956 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq5cp\" (UniqueName: \"kubernetes.io/projected/a46be1f3-50fc-45b5-a480-98d9763db69d-kube-api-access-lq5cp\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.279028 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.366266 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-vb9qx" podUID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" containerName="console" containerID="cri-o://32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b" gracePeriod=15 Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.380770 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.380895 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.381420 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.381383 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.381429 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq5cp\" (UniqueName: \"kubernetes.io/projected/a46be1f3-50fc-45b5-a480-98d9763db69d-kube-api-access-lq5cp\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.400388 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq5cp\" (UniqueName: \"kubernetes.io/projected/a46be1f3-50fc-45b5-a480-98d9763db69d-kube-api-access-lq5cp\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.517906 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.695088 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vb9qx_270031fa-3d83-4edf-bb5d-19ce9e1a693d/console/0.log" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.695351 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.702806 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg"] Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.788289 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t46vq\" (UniqueName: \"kubernetes.io/projected/270031fa-3d83-4edf-bb5d-19ce9e1a693d-kube-api-access-t46vq\") pod \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.788387 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-trusted-ca-bundle\") pod \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.788525 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-config\") pod \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.788573 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-oauth-config\") pod \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.788634 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-serving-cert\") pod \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.788694 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-oauth-serving-cert\") pod \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.788758 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-service-ca\") pod \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\" (UID: \"270031fa-3d83-4edf-bb5d-19ce9e1a693d\") " Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.789482 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-config" (OuterVolumeSpecName: "console-config") pod "270031fa-3d83-4edf-bb5d-19ce9e1a693d" (UID: "270031fa-3d83-4edf-bb5d-19ce9e1a693d"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.790026 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "270031fa-3d83-4edf-bb5d-19ce9e1a693d" (UID: "270031fa-3d83-4edf-bb5d-19ce9e1a693d"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.790179 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-service-ca" (OuterVolumeSpecName: "service-ca") pod "270031fa-3d83-4edf-bb5d-19ce9e1a693d" (UID: "270031fa-3d83-4edf-bb5d-19ce9e1a693d"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.790218 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "270031fa-3d83-4edf-bb5d-19ce9e1a693d" (UID: "270031fa-3d83-4edf-bb5d-19ce9e1a693d"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.795475 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "270031fa-3d83-4edf-bb5d-19ce9e1a693d" (UID: "270031fa-3d83-4edf-bb5d-19ce9e1a693d"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.795989 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "270031fa-3d83-4edf-bb5d-19ce9e1a693d" (UID: "270031fa-3d83-4edf-bb5d-19ce9e1a693d"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.797596 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270031fa-3d83-4edf-bb5d-19ce9e1a693d-kube-api-access-t46vq" (OuterVolumeSpecName: "kube-api-access-t46vq") pod "270031fa-3d83-4edf-bb5d-19ce9e1a693d" (UID: "270031fa-3d83-4edf-bb5d-19ce9e1a693d"). InnerVolumeSpecName "kube-api-access-t46vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.890544 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t46vq\" (UniqueName: \"kubernetes.io/projected/270031fa-3d83-4edf-bb5d-19ce9e1a693d-kube-api-access-t46vq\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.890571 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.890580 4926 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.890588 4926 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.890597 4926 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/270031fa-3d83-4edf-bb5d-19ce9e1a693d-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.890606 4926 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:29 crc kubenswrapper[4926]: I0312 18:17:29.890614 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/270031fa-3d83-4edf-bb5d-19ce9e1a693d-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.436700 4926 generic.go:334] "Generic (PLEG): container finished" podID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerID="990a6bbc9787734ec00d8b44f262ac34b09786254030da04e8d02b2392609120" exitCode=0 Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.436789 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" event={"ID":"a46be1f3-50fc-45b5-a480-98d9763db69d","Type":"ContainerDied","Data":"990a6bbc9787734ec00d8b44f262ac34b09786254030da04e8d02b2392609120"} Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.437049 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" event={"ID":"a46be1f3-50fc-45b5-a480-98d9763db69d","Type":"ContainerStarted","Data":"2f45bee694048e411034283204782edde4d63f3b997a31a54e77437bd55ef983"} Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.440519 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-vb9qx_270031fa-3d83-4edf-bb5d-19ce9e1a693d/console/0.log" Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.440617 4926 generic.go:334] "Generic (PLEG): container finished" podID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" containerID="32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b" exitCode=2 Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.440665 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vb9qx" event={"ID":"270031fa-3d83-4edf-bb5d-19ce9e1a693d","Type":"ContainerDied","Data":"32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b"} Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.440704 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vb9qx" event={"ID":"270031fa-3d83-4edf-bb5d-19ce9e1a693d","Type":"ContainerDied","Data":"8b8722347bac291c780b1fdd439b9d8743876830528929d7bd120c0953d14111"} Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.440733 4926 scope.go:117] "RemoveContainer" containerID="32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b" Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.440921 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vb9qx" Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.441026 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.500283 4926 scope.go:117] "RemoveContainer" containerID="32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b" Mar 12 18:17:30 crc kubenswrapper[4926]: E0312 18:17:30.501007 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b\": container with ID starting with 32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b not found: ID does not exist" containerID="32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b" Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.501061 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b"} err="failed to get container status \"32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b\": rpc error: code = NotFound desc = could not find container \"32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b\": container with ID starting with 32bb48f2aa4bede9bbf0c7b620e2f860f26e2deae6fde3c60c9a415822d1b25b not found: ID does not exist" Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.510914 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-vb9qx"] Mar 12 18:17:30 crc kubenswrapper[4926]: I0312 18:17:30.510971 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-vb9qx"] Mar 12 18:17:32 crc kubenswrapper[4926]: I0312 18:17:32.590462 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" path="/var/lib/kubelet/pods/270031fa-3d83-4edf-bb5d-19ce9e1a693d/volumes" Mar 12 18:17:32 crc kubenswrapper[4926]: I0312 18:17:32.592196 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" event={"ID":"a46be1f3-50fc-45b5-a480-98d9763db69d","Type":"ContainerStarted","Data":"ccb2939e93467122e809759af18c473c961879e45f9a02cfc6a36f1cb08e16c7"} Mar 12 18:17:33 crc kubenswrapper[4926]: I0312 18:17:33.605752 4926 generic.go:334] "Generic (PLEG): container finished" podID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerID="ccb2939e93467122e809759af18c473c961879e45f9a02cfc6a36f1cb08e16c7" exitCode=0 Mar 12 18:17:33 crc kubenswrapper[4926]: I0312 18:17:33.605833 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" event={"ID":"a46be1f3-50fc-45b5-a480-98d9763db69d","Type":"ContainerDied","Data":"ccb2939e93467122e809759af18c473c961879e45f9a02cfc6a36f1cb08e16c7"} Mar 12 18:17:34 crc kubenswrapper[4926]: I0312 18:17:34.616329 4926 generic.go:334] "Generic (PLEG): container finished" podID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerID="2c66fe18bef8f2684627fbbbebbb90bceda864ef27b9acf48fed31542186e3f5" exitCode=0 Mar 12 18:17:34 crc kubenswrapper[4926]: I0312 18:17:34.616581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" event={"ID":"a46be1f3-50fc-45b5-a480-98d9763db69d","Type":"ContainerDied","Data":"2c66fe18bef8f2684627fbbbebbb90bceda864ef27b9acf48fed31542186e3f5"} Mar 12 18:17:35 crc kubenswrapper[4926]: I0312 18:17:35.933528 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:35 crc kubenswrapper[4926]: I0312 18:17:35.974313 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-util\") pod \"a46be1f3-50fc-45b5-a480-98d9763db69d\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " Mar 12 18:17:35 crc kubenswrapper[4926]: I0312 18:17:35.974645 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq5cp\" (UniqueName: \"kubernetes.io/projected/a46be1f3-50fc-45b5-a480-98d9763db69d-kube-api-access-lq5cp\") pod \"a46be1f3-50fc-45b5-a480-98d9763db69d\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " Mar 12 18:17:35 crc kubenswrapper[4926]: I0312 18:17:35.974737 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-bundle\") pod \"a46be1f3-50fc-45b5-a480-98d9763db69d\" (UID: \"a46be1f3-50fc-45b5-a480-98d9763db69d\") " Mar 12 18:17:35 crc kubenswrapper[4926]: I0312 18:17:35.976645 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-bundle" (OuterVolumeSpecName: "bundle") pod "a46be1f3-50fc-45b5-a480-98d9763db69d" (UID: "a46be1f3-50fc-45b5-a480-98d9763db69d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:17:35 crc kubenswrapper[4926]: I0312 18:17:35.980878 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a46be1f3-50fc-45b5-a480-98d9763db69d-kube-api-access-lq5cp" (OuterVolumeSpecName: "kube-api-access-lq5cp") pod "a46be1f3-50fc-45b5-a480-98d9763db69d" (UID: "a46be1f3-50fc-45b5-a480-98d9763db69d"). InnerVolumeSpecName "kube-api-access-lq5cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:17:36 crc kubenswrapper[4926]: I0312 18:17:36.000122 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-util" (OuterVolumeSpecName: "util") pod "a46be1f3-50fc-45b5-a480-98d9763db69d" (UID: "a46be1f3-50fc-45b5-a480-98d9763db69d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:17:36 crc kubenswrapper[4926]: I0312 18:17:36.076657 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:36 crc kubenswrapper[4926]: I0312 18:17:36.076714 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a46be1f3-50fc-45b5-a480-98d9763db69d-util\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:36 crc kubenswrapper[4926]: I0312 18:17:36.076726 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq5cp\" (UniqueName: \"kubernetes.io/projected/a46be1f3-50fc-45b5-a480-98d9763db69d-kube-api-access-lq5cp\") on node \"crc\" DevicePath \"\"" Mar 12 18:17:36 crc kubenswrapper[4926]: I0312 18:17:36.635646 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" event={"ID":"a46be1f3-50fc-45b5-a480-98d9763db69d","Type":"ContainerDied","Data":"2f45bee694048e411034283204782edde4d63f3b997a31a54e77437bd55ef983"} Mar 12 18:17:36 crc kubenswrapper[4926]: I0312 18:17:36.635682 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f45bee694048e411034283204782edde4d63f3b997a31a54e77437bd55ef983" Mar 12 18:17:36 crc kubenswrapper[4926]: I0312 18:17:36.635788 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.638257 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8"] Mar 12 18:17:44 crc kubenswrapper[4926]: E0312 18:17:44.639093 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" containerName="console" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.639108 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" containerName="console" Mar 12 18:17:44 crc kubenswrapper[4926]: E0312 18:17:44.639120 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerName="pull" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.639127 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerName="pull" Mar 12 18:17:44 crc kubenswrapper[4926]: E0312 18:17:44.639136 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerName="extract" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.639145 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerName="extract" Mar 12 18:17:44 crc kubenswrapper[4926]: E0312 18:17:44.639158 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerName="util" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.639165 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerName="util" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.639281 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="270031fa-3d83-4edf-bb5d-19ce9e1a693d" containerName="console" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.639294 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a46be1f3-50fc-45b5-a480-98d9763db69d" containerName="extract" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.639816 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.642100 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.642228 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.642457 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-9rnfq" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.643255 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.643792 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.702359 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a847a81-61ae-42e3-9866-c25b68fd77cb-webhook-cert\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.702452 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfj4m\" (UniqueName: \"kubernetes.io/projected/8a847a81-61ae-42e3-9866-c25b68fd77cb-kube-api-access-jfj4m\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.702503 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a847a81-61ae-42e3-9866-c25b68fd77cb-apiservice-cert\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.705122 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8"] Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.803889 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a847a81-61ae-42e3-9866-c25b68fd77cb-webhook-cert\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.804279 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfj4m\" (UniqueName: \"kubernetes.io/projected/8a847a81-61ae-42e3-9866-c25b68fd77cb-kube-api-access-jfj4m\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.804332 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a847a81-61ae-42e3-9866-c25b68fd77cb-apiservice-cert\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.814933 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a847a81-61ae-42e3-9866-c25b68fd77cb-apiservice-cert\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.824277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a847a81-61ae-42e3-9866-c25b68fd77cb-webhook-cert\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.828896 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfj4m\" (UniqueName: \"kubernetes.io/projected/8a847a81-61ae-42e3-9866-c25b68fd77cb-kube-api-access-jfj4m\") pod \"metallb-operator-controller-manager-7cf578c5b8-z4gn8\" (UID: \"8a847a81-61ae-42e3-9866-c25b68fd77cb\") " pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.959175 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.967173 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58"] Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.967906 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.970160 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.970317 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 12 18:17:44 crc kubenswrapper[4926]: I0312 18:17:44.970548 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-gjjw5" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.020991 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mmds\" (UniqueName: \"kubernetes.io/projected/4d6f4326-2022-436b-9523-383aae3fd5cd-kube-api-access-6mmds\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.021086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4d6f4326-2022-436b-9523-383aae3fd5cd-apiservice-cert\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.021149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4d6f4326-2022-436b-9523-383aae3fd5cd-webhook-cert\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.061118 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58"] Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.122211 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mmds\" (UniqueName: \"kubernetes.io/projected/4d6f4326-2022-436b-9523-383aae3fd5cd-kube-api-access-6mmds\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.122297 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4d6f4326-2022-436b-9523-383aae3fd5cd-apiservice-cert\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.122337 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4d6f4326-2022-436b-9523-383aae3fd5cd-webhook-cert\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.131719 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4d6f4326-2022-436b-9523-383aae3fd5cd-webhook-cert\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.137356 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4d6f4326-2022-436b-9523-383aae3fd5cd-apiservice-cert\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.162040 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mmds\" (UniqueName: \"kubernetes.io/projected/4d6f4326-2022-436b-9523-383aae3fd5cd-kube-api-access-6mmds\") pod \"metallb-operator-webhook-server-79dbc878dc-w7f58\" (UID: \"4d6f4326-2022-436b-9523-383aae3fd5cd\") " pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.235625 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8"] Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.343711 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.637265 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58"] Mar 12 18:17:45 crc kubenswrapper[4926]: W0312 18:17:45.644299 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d6f4326_2022_436b_9523_383aae3fd5cd.slice/crio-61f6d9cd37848b3546c922f5be12b1797c041f0cdc0df201f6726cd850cb1afd WatchSource:0}: Error finding container 61f6d9cd37848b3546c922f5be12b1797c041f0cdc0df201f6726cd850cb1afd: Status 404 returned error can't find the container with id 61f6d9cd37848b3546c922f5be12b1797c041f0cdc0df201f6726cd850cb1afd Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.685593 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" event={"ID":"8a847a81-61ae-42e3-9866-c25b68fd77cb","Type":"ContainerStarted","Data":"9beb4ee77df9449d491ad0358f61dadbb88c3ea4e7c5232740b2ec2dc5613459"} Mar 12 18:17:45 crc kubenswrapper[4926]: I0312 18:17:45.687088 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" event={"ID":"4d6f4326-2022-436b-9523-383aae3fd5cd","Type":"ContainerStarted","Data":"61f6d9cd37848b3546c922f5be12b1797c041f0cdc0df201f6726cd850cb1afd"} Mar 12 18:17:51 crc kubenswrapper[4926]: I0312 18:17:51.724145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" event={"ID":"8a847a81-61ae-42e3-9866-c25b68fd77cb","Type":"ContainerStarted","Data":"c44c4510181cfa327e64fc978bdcabc5a11fee48eceb0d971b89a6a7b1e29bcc"} Mar 12 18:17:51 crc kubenswrapper[4926]: I0312 18:17:51.725960 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:17:51 crc kubenswrapper[4926]: I0312 18:17:51.728227 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" event={"ID":"4d6f4326-2022-436b-9523-383aae3fd5cd","Type":"ContainerStarted","Data":"e591d599c1b024d8b866236eb0fb8a1794e0bfc252c1693353dac27a1e0a2644"} Mar 12 18:17:51 crc kubenswrapper[4926]: I0312 18:17:51.728360 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:17:51 crc kubenswrapper[4926]: I0312 18:17:51.785100 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" podStartSLOduration=1.9878568699999999 podStartE2EDuration="7.785071173s" podCreationTimestamp="2026-03-12 18:17:44 +0000 UTC" firstStartedPulling="2026-03-12 18:17:45.245625298 +0000 UTC m=+905.614251631" lastFinishedPulling="2026-03-12 18:17:51.042839601 +0000 UTC m=+911.411465934" observedRunningTime="2026-03-12 18:17:51.771932534 +0000 UTC m=+912.140558877" watchObservedRunningTime="2026-03-12 18:17:51.785071173 +0000 UTC m=+912.153697516" Mar 12 18:17:51 crc kubenswrapper[4926]: I0312 18:17:51.821229 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" podStartSLOduration=2.396985769 podStartE2EDuration="7.821212458s" podCreationTimestamp="2026-03-12 18:17:44 +0000 UTC" firstStartedPulling="2026-03-12 18:17:45.647224432 +0000 UTC m=+906.015850755" lastFinishedPulling="2026-03-12 18:17:51.071451111 +0000 UTC m=+911.440077444" observedRunningTime="2026-03-12 18:17:51.820286749 +0000 UTC m=+912.188913122" watchObservedRunningTime="2026-03-12 18:17:51.821212458 +0000 UTC m=+912.189838791" Mar 12 18:17:56 crc kubenswrapper[4926]: I0312 18:17:56.817235 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:17:56 crc kubenswrapper[4926]: I0312 18:17:56.817774 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:17:56 crc kubenswrapper[4926]: I0312 18:17:56.817818 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:17:56 crc kubenswrapper[4926]: I0312 18:17:56.818325 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a397bef079b1410b3294983dad25ada9109b1a0eac364c78c0ff4aeeccdf38ed"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:17:56 crc kubenswrapper[4926]: I0312 18:17:56.818371 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://a397bef079b1410b3294983dad25ada9109b1a0eac364c78c0ff4aeeccdf38ed" gracePeriod=600 Mar 12 18:17:57 crc kubenswrapper[4926]: I0312 18:17:57.765701 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="a397bef079b1410b3294983dad25ada9109b1a0eac364c78c0ff4aeeccdf38ed" exitCode=0 Mar 12 18:17:57 crc kubenswrapper[4926]: I0312 18:17:57.765788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"a397bef079b1410b3294983dad25ada9109b1a0eac364c78c0ff4aeeccdf38ed"} Mar 12 18:17:57 crc kubenswrapper[4926]: I0312 18:17:57.765974 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"10c4816f4e2fc4ce2bc2183a633d9bc53980639515bfce0cf198e862b133fadb"} Mar 12 18:17:57 crc kubenswrapper[4926]: I0312 18:17:57.765994 4926 scope.go:117] "RemoveContainer" containerID="869151c9e3071e8f72a54f977df4fbec55cdf81d3f75158a024a468d8d420c6b" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.121570 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555658-gfgnc"] Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.122881 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.131905 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555658-gfgnc"] Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.132693 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.134510 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.136779 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.249730 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnhr6\" (UniqueName: \"kubernetes.io/projected/4b727d6f-6c9d-44ad-8594-accc9d2c4ed6-kube-api-access-hnhr6\") pod \"auto-csr-approver-29555658-gfgnc\" (UID: \"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6\") " pod="openshift-infra/auto-csr-approver-29555658-gfgnc" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.350724 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnhr6\" (UniqueName: \"kubernetes.io/projected/4b727d6f-6c9d-44ad-8594-accc9d2c4ed6-kube-api-access-hnhr6\") pod \"auto-csr-approver-29555658-gfgnc\" (UID: \"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6\") " pod="openshift-infra/auto-csr-approver-29555658-gfgnc" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.379751 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnhr6\" (UniqueName: \"kubernetes.io/projected/4b727d6f-6c9d-44ad-8594-accc9d2c4ed6-kube-api-access-hnhr6\") pod \"auto-csr-approver-29555658-gfgnc\" (UID: \"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6\") " pod="openshift-infra/auto-csr-approver-29555658-gfgnc" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.440035 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.657726 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555658-gfgnc"] Mar 12 18:18:00 crc kubenswrapper[4926]: I0312 18:18:00.786323 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" event={"ID":"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6","Type":"ContainerStarted","Data":"be8d63d1d24bbb247fb5937d6f970c255d35cfad381121636213a2bba32500dc"} Mar 12 18:18:02 crc kubenswrapper[4926]: I0312 18:18:02.806459 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" event={"ID":"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6","Type":"ContainerStarted","Data":"287cfd2e2e77a9b487765ae984d056f483b704272f97333956b53dc7f101b02d"} Mar 12 18:18:02 crc kubenswrapper[4926]: I0312 18:18:02.822072 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" podStartSLOduration=1.149782568 podStartE2EDuration="2.822055096s" podCreationTimestamp="2026-03-12 18:18:00 +0000 UTC" firstStartedPulling="2026-03-12 18:18:00.664755629 +0000 UTC m=+921.033381972" lastFinishedPulling="2026-03-12 18:18:02.337028167 +0000 UTC m=+922.705654500" observedRunningTime="2026-03-12 18:18:02.818156726 +0000 UTC m=+923.186783059" watchObservedRunningTime="2026-03-12 18:18:02.822055096 +0000 UTC m=+923.190681429" Mar 12 18:18:03 crc kubenswrapper[4926]: I0312 18:18:03.815334 4926 generic.go:334] "Generic (PLEG): container finished" podID="4b727d6f-6c9d-44ad-8594-accc9d2c4ed6" containerID="287cfd2e2e77a9b487765ae984d056f483b704272f97333956b53dc7f101b02d" exitCode=0 Mar 12 18:18:03 crc kubenswrapper[4926]: I0312 18:18:03.815401 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" event={"ID":"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6","Type":"ContainerDied","Data":"287cfd2e2e77a9b487765ae984d056f483b704272f97333956b53dc7f101b02d"} Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.133021 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.214847 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnhr6\" (UniqueName: \"kubernetes.io/projected/4b727d6f-6c9d-44ad-8594-accc9d2c4ed6-kube-api-access-hnhr6\") pod \"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6\" (UID: \"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6\") " Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.225668 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b727d6f-6c9d-44ad-8594-accc9d2c4ed6-kube-api-access-hnhr6" (OuterVolumeSpecName: "kube-api-access-hnhr6") pod "4b727d6f-6c9d-44ad-8594-accc9d2c4ed6" (UID: "4b727d6f-6c9d-44ad-8594-accc9d2c4ed6"). InnerVolumeSpecName "kube-api-access-hnhr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.316431 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnhr6\" (UniqueName: \"kubernetes.io/projected/4b727d6f-6c9d-44ad-8594-accc9d2c4ed6-kube-api-access-hnhr6\") on node \"crc\" DevicePath \"\"" Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.348406 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-79dbc878dc-w7f58" Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.830340 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" event={"ID":"4b727d6f-6c9d-44ad-8594-accc9d2c4ed6","Type":"ContainerDied","Data":"be8d63d1d24bbb247fb5937d6f970c255d35cfad381121636213a2bba32500dc"} Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.830383 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be8d63d1d24bbb247fb5937d6f970c255d35cfad381121636213a2bba32500dc" Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.830461 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555658-gfgnc" Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.883625 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555652-mrcx4"] Mar 12 18:18:05 crc kubenswrapper[4926]: I0312 18:18:05.890901 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555652-mrcx4"] Mar 12 18:18:06 crc kubenswrapper[4926]: I0312 18:18:06.497970 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a135018f-2c21-4678-a0fc-9d6b62dda2d6" path="/var/lib/kubelet/pods/a135018f-2c21-4678-a0fc-9d6b62dda2d6/volumes" Mar 12 18:18:24 crc kubenswrapper[4926]: I0312 18:18:24.962673 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7cf578c5b8-z4gn8" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.694485 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d"] Mar 12 18:18:25 crc kubenswrapper[4926]: E0312 18:18:25.694857 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b727d6f-6c9d-44ad-8594-accc9d2c4ed6" containerName="oc" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.694882 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b727d6f-6c9d-44ad-8594-accc9d2c4ed6" containerName="oc" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.695058 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b727d6f-6c9d-44ad-8594-accc9d2c4ed6" containerName="oc" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.695687 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.699029 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.699101 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-2thch" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.701484 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-9p72p"] Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.704350 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.706018 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.706252 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.710259 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d"] Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.774156 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6bt9q"] Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.775182 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.777510 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.777716 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.777830 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.777943 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-t48cm" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794663 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-reloader\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794724 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-sockets\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794752 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjp96\" (UniqueName: \"kubernetes.io/projected/036c2795-2942-4cc8-9a91-6cc48cbe7521-kube-api-access-zjp96\") pod \"frr-k8s-webhook-server-bcc4b6f68-j2n6d\" (UID: \"036c2795-2942-4cc8-9a91-6cc48cbe7521\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794777 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-startup\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794799 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-conf\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794821 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/036c2795-2942-4cc8-9a91-6cc48cbe7521-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-j2n6d\" (UID: \"036c2795-2942-4cc8-9a91-6cc48cbe7521\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794839 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794911 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzhdm\" (UniqueName: \"kubernetes.io/projected/53da3fff-e3f4-4b9d-a887-f5a28f986107-kube-api-access-zzhdm\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.794933 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics-certs\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.797362 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-58dtw"] Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.798735 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.808072 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.826828 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-58dtw"] Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.895990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-sockets\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896041 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjp96\" (UniqueName: \"kubernetes.io/projected/036c2795-2942-4cc8-9a91-6cc48cbe7521-kube-api-access-zjp96\") pod \"frr-k8s-webhook-server-bcc4b6f68-j2n6d\" (UID: \"036c2795-2942-4cc8-9a91-6cc48cbe7521\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896060 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-startup\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896078 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-conf\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896097 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896113 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/036c2795-2942-4cc8-9a91-6cc48cbe7521-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-j2n6d\" (UID: \"036c2795-2942-4cc8-9a91-6cc48cbe7521\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896134 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-cert\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896163 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896181 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxzrt\" (UniqueName: \"kubernetes.io/projected/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-kube-api-access-lxzrt\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896209 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-metrics-certs\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896286 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84cq2\" (UniqueName: \"kubernetes.io/projected/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-kube-api-access-84cq2\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896305 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzhdm\" (UniqueName: \"kubernetes.io/projected/53da3fff-e3f4-4b9d-a887-f5a28f986107-kube-api-access-zzhdm\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896320 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics-certs\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896347 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-metallb-excludel2\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896370 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-metrics-certs\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: E0312 18:18:25.896380 4926 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 12 18:18:25 crc kubenswrapper[4926]: E0312 18:18:25.896484 4926 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Mar 12 18:18:25 crc kubenswrapper[4926]: E0312 18:18:25.896504 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/036c2795-2942-4cc8-9a91-6cc48cbe7521-cert podName:036c2795-2942-4cc8-9a91-6cc48cbe7521 nodeName:}" failed. No retries permitted until 2026-03-12 18:18:26.396449129 +0000 UTC m=+946.765075472 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/036c2795-2942-4cc8-9a91-6cc48cbe7521-cert") pod "frr-k8s-webhook-server-bcc4b6f68-j2n6d" (UID: "036c2795-2942-4cc8-9a91-6cc48cbe7521") : secret "frr-k8s-webhook-server-cert" not found Mar 12 18:18:25 crc kubenswrapper[4926]: E0312 18:18:25.896535 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics-certs podName:53da3fff-e3f4-4b9d-a887-f5a28f986107 nodeName:}" failed. No retries permitted until 2026-03-12 18:18:26.396514781 +0000 UTC m=+946.765141114 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics-certs") pod "frr-k8s-9p72p" (UID: "53da3fff-e3f4-4b9d-a887-f5a28f986107") : secret "frr-k8s-certs-secret" not found Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896492 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-sockets\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896551 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-reloader\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896621 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-conf\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896697 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.896826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/53da3fff-e3f4-4b9d-a887-f5a28f986107-reloader\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.897131 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/53da3fff-e3f4-4b9d-a887-f5a28f986107-frr-startup\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.917337 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzhdm\" (UniqueName: \"kubernetes.io/projected/53da3fff-e3f4-4b9d-a887-f5a28f986107-kube-api-access-zzhdm\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.918094 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjp96\" (UniqueName: \"kubernetes.io/projected/036c2795-2942-4cc8-9a91-6cc48cbe7521-kube-api-access-zjp96\") pod \"frr-k8s-webhook-server-bcc4b6f68-j2n6d\" (UID: \"036c2795-2942-4cc8-9a91-6cc48cbe7521\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.998342 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84cq2\" (UniqueName: \"kubernetes.io/projected/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-kube-api-access-84cq2\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.998415 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-metallb-excludel2\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.998462 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-metrics-certs\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.998531 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-cert\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.998559 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.998580 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxzrt\" (UniqueName: \"kubernetes.io/projected/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-kube-api-access-lxzrt\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.998613 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-metrics-certs\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:25 crc kubenswrapper[4926]: E0312 18:18:25.998819 4926 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 12 18:18:25 crc kubenswrapper[4926]: E0312 18:18:25.998910 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist podName:d10f0ca7-6fc4-4e6a-815c-ad5a1db16350 nodeName:}" failed. No retries permitted until 2026-03-12 18:18:26.498888876 +0000 UTC m=+946.867515219 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist") pod "speaker-6bt9q" (UID: "d10f0ca7-6fc4-4e6a-815c-ad5a1db16350") : secret "metallb-memberlist" not found Mar 12 18:18:25 crc kubenswrapper[4926]: I0312 18:18:25.999279 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-metallb-excludel2\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.001878 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-metrics-certs\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.002138 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-metrics-certs\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.002318 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-cert\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.016199 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84cq2\" (UniqueName: \"kubernetes.io/projected/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-kube-api-access-84cq2\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.026959 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxzrt\" (UniqueName: \"kubernetes.io/projected/16ea4f33-38fb-42c7-9c85-67c443f0b3a4-kube-api-access-lxzrt\") pod \"controller-7bb4cc7c98-58dtw\" (UID: \"16ea4f33-38fb-42c7-9c85-67c443f0b3a4\") " pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.128177 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.403399 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics-certs\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.403811 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/036c2795-2942-4cc8-9a91-6cc48cbe7521-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-j2n6d\" (UID: \"036c2795-2942-4cc8-9a91-6cc48cbe7521\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.409003 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/53da3fff-e3f4-4b9d-a887-f5a28f986107-metrics-certs\") pod \"frr-k8s-9p72p\" (UID: \"53da3fff-e3f4-4b9d-a887-f5a28f986107\") " pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.409576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/036c2795-2942-4cc8-9a91-6cc48cbe7521-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-j2n6d\" (UID: \"036c2795-2942-4cc8-9a91-6cc48cbe7521\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.505535 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:26 crc kubenswrapper[4926]: E0312 18:18:26.505684 4926 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 12 18:18:26 crc kubenswrapper[4926]: E0312 18:18:26.505737 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist podName:d10f0ca7-6fc4-4e6a-815c-ad5a1db16350 nodeName:}" failed. No retries permitted until 2026-03-12 18:18:27.505720805 +0000 UTC m=+947.874347138 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist") pod "speaker-6bt9q" (UID: "d10f0ca7-6fc4-4e6a-815c-ad5a1db16350") : secret "metallb-memberlist" not found Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.537710 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-58dtw"] Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.622724 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.638880 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.861401 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d"] Mar 12 18:18:26 crc kubenswrapper[4926]: W0312 18:18:26.864909 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod036c2795_2942_4cc8_9a91_6cc48cbe7521.slice/crio-af71a557196e989147aaf222c4407e462d264b77648a93a33cb0b35b3d989008 WatchSource:0}: Error finding container af71a557196e989147aaf222c4407e462d264b77648a93a33cb0b35b3d989008: Status 404 returned error can't find the container with id af71a557196e989147aaf222c4407e462d264b77648a93a33cb0b35b3d989008 Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.975012 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerStarted","Data":"7bc9d3477ec1337148248500fa8f077509e6af660b9e36f31ca9c796ca21db3c"} Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.976154 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" event={"ID":"036c2795-2942-4cc8-9a91-6cc48cbe7521","Type":"ContainerStarted","Data":"af71a557196e989147aaf222c4407e462d264b77648a93a33cb0b35b3d989008"} Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.978325 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-58dtw" event={"ID":"16ea4f33-38fb-42c7-9c85-67c443f0b3a4","Type":"ContainerStarted","Data":"03c9a6572f409b14b5d79d225c8d698d54ce5f2ad0d5248d916c606a8887b589"} Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.978369 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-58dtw" event={"ID":"16ea4f33-38fb-42c7-9c85-67c443f0b3a4","Type":"ContainerStarted","Data":"a85373b45dbbe732c0eba8edc211072a173bb3fa500977a7dc241c0dd6a3a6ca"} Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.978380 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-58dtw" event={"ID":"16ea4f33-38fb-42c7-9c85-67c443f0b3a4","Type":"ContainerStarted","Data":"51a47c9019e7bdcbf4b49d7c372cbd35b4692e9ef8de14fcd80d7f16b9e28168"} Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.978471 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:26 crc kubenswrapper[4926]: I0312 18:18:26.994816 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-58dtw" podStartSLOduration=1.994801211 podStartE2EDuration="1.994801211s" podCreationTimestamp="2026-03-12 18:18:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:18:26.99059688 +0000 UTC m=+947.359223213" watchObservedRunningTime="2026-03-12 18:18:26.994801211 +0000 UTC m=+947.363427544" Mar 12 18:18:27 crc kubenswrapper[4926]: I0312 18:18:27.522684 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:27 crc kubenswrapper[4926]: I0312 18:18:27.529625 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d10f0ca7-6fc4-4e6a-815c-ad5a1db16350-memberlist\") pod \"speaker-6bt9q\" (UID: \"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350\") " pod="metallb-system/speaker-6bt9q" Mar 12 18:18:27 crc kubenswrapper[4926]: I0312 18:18:27.591010 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6bt9q" Mar 12 18:18:27 crc kubenswrapper[4926]: I0312 18:18:27.986728 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6bt9q" event={"ID":"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350","Type":"ContainerStarted","Data":"479de8f9c60ff091edbc68c3a1ec82bf8519f9a0c67baefdb7b6e478740c5685"} Mar 12 18:18:27 crc kubenswrapper[4926]: I0312 18:18:27.986797 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6bt9q" event={"ID":"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350","Type":"ContainerStarted","Data":"cffa8917727a15408f12c2bf94d6fd73ecf0e805a71196c59b24ca6f44afb9b8"} Mar 12 18:18:28 crc kubenswrapper[4926]: I0312 18:18:28.994348 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6bt9q" event={"ID":"d10f0ca7-6fc4-4e6a-815c-ad5a1db16350","Type":"ContainerStarted","Data":"76ce3956f3b260add7de62d4cd4e7373ef3bf207020bbea3a061106566998243"} Mar 12 18:18:28 crc kubenswrapper[4926]: I0312 18:18:28.994619 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6bt9q" Mar 12 18:18:29 crc kubenswrapper[4926]: I0312 18:18:29.020791 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6bt9q" podStartSLOduration=4.020768243 podStartE2EDuration="4.020768243s" podCreationTimestamp="2026-03-12 18:18:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:18:29.016134919 +0000 UTC m=+949.384761252" watchObservedRunningTime="2026-03-12 18:18:29.020768243 +0000 UTC m=+949.389394576" Mar 12 18:18:36 crc kubenswrapper[4926]: I0312 18:18:36.133160 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-58dtw" Mar 12 18:18:37 crc kubenswrapper[4926]: I0312 18:18:37.090672 4926 generic.go:334] "Generic (PLEG): container finished" podID="53da3fff-e3f4-4b9d-a887-f5a28f986107" containerID="764bc52e4da56ad2ecb50ab328be575e043ec37d17dee06fbb12ab36d7af44c8" exitCode=0 Mar 12 18:18:37 crc kubenswrapper[4926]: I0312 18:18:37.090751 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerDied","Data":"764bc52e4da56ad2ecb50ab328be575e043ec37d17dee06fbb12ab36d7af44c8"} Mar 12 18:18:37 crc kubenswrapper[4926]: I0312 18:18:37.092510 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" event={"ID":"036c2795-2942-4cc8-9a91-6cc48cbe7521","Type":"ContainerStarted","Data":"c699b6568bed00fc040b1d65e6af418f4fdbc86390fce722ce49b05c81835a8b"} Mar 12 18:18:37 crc kubenswrapper[4926]: I0312 18:18:37.092713 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:37 crc kubenswrapper[4926]: I0312 18:18:37.154999 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" podStartSLOduration=3.092244015 podStartE2EDuration="12.154981505s" podCreationTimestamp="2026-03-12 18:18:25 +0000 UTC" firstStartedPulling="2026-03-12 18:18:26.867938954 +0000 UTC m=+947.236565287" lastFinishedPulling="2026-03-12 18:18:35.930676434 +0000 UTC m=+956.299302777" observedRunningTime="2026-03-12 18:18:37.153823379 +0000 UTC m=+957.522449712" watchObservedRunningTime="2026-03-12 18:18:37.154981505 +0000 UTC m=+957.523607838" Mar 12 18:18:37 crc kubenswrapper[4926]: I0312 18:18:37.596012 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6bt9q" Mar 12 18:18:38 crc kubenswrapper[4926]: I0312 18:18:38.103804 4926 generic.go:334] "Generic (PLEG): container finished" podID="53da3fff-e3f4-4b9d-a887-f5a28f986107" containerID="060ed7cddeb7f6b2a72d027d9728b1971913869840a312f2fc04fea89d0d6142" exitCode=0 Mar 12 18:18:38 crc kubenswrapper[4926]: I0312 18:18:38.103906 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerDied","Data":"060ed7cddeb7f6b2a72d027d9728b1971913869840a312f2fc04fea89d0d6142"} Mar 12 18:18:39 crc kubenswrapper[4926]: I0312 18:18:39.111942 4926 generic.go:334] "Generic (PLEG): container finished" podID="53da3fff-e3f4-4b9d-a887-f5a28f986107" containerID="b36c63abfe5b08cdc3a148b062643ebd6e582fffc5989f4a952be3fd61137963" exitCode=0 Mar 12 18:18:39 crc kubenswrapper[4926]: I0312 18:18:39.111983 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerDied","Data":"b36c63abfe5b08cdc3a148b062643ebd6e582fffc5989f4a952be3fd61137963"} Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.122731 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerStarted","Data":"bcfb93ea45aae8696c20fa8649694239781a5bd2ecd56d2e8e91e41a85b4384d"} Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.123056 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerStarted","Data":"1d8f0804408b03049e96fc9415277c1c2a291f83e8b7cd21c0fe8809bc68868e"} Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.123069 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerStarted","Data":"204cb34230ede5855ffb23a7568e4a8fe97836d9b043d4c2ce1417955a626e5e"} Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.123078 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerStarted","Data":"42dd16964c7e152638199f8c806180078e0b4b5b008d51c7b01231e9f537b3c5"} Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.123087 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerStarted","Data":"484ec29a538864455de1aa755032e7b37e46e22e97117bd712533afa9a92e162"} Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.265120 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-vj2fp"] Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.265982 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vj2fp" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.269294 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-sqgdz" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.269660 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.270278 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.278790 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vj2fp"] Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.411112 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-js5nb\" (UniqueName: \"kubernetes.io/projected/c7718f63-5b8f-4137-8a80-f8793ce3e60f-kube-api-access-js5nb\") pod \"openstack-operator-index-vj2fp\" (UID: \"c7718f63-5b8f-4137-8a80-f8793ce3e60f\") " pod="openstack-operators/openstack-operator-index-vj2fp" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.512574 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-js5nb\" (UniqueName: \"kubernetes.io/projected/c7718f63-5b8f-4137-8a80-f8793ce3e60f-kube-api-access-js5nb\") pod \"openstack-operator-index-vj2fp\" (UID: \"c7718f63-5b8f-4137-8a80-f8793ce3e60f\") " pod="openstack-operators/openstack-operator-index-vj2fp" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.525471 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.535844 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.561815 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-js5nb\" (UniqueName: \"kubernetes.io/projected/c7718f63-5b8f-4137-8a80-f8793ce3e60f-kube-api-access-js5nb\") pod \"openstack-operator-index-vj2fp\" (UID: \"c7718f63-5b8f-4137-8a80-f8793ce3e60f\") " pod="openstack-operators/openstack-operator-index-vj2fp" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.640325 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-sqgdz" Mar 12 18:18:40 crc kubenswrapper[4926]: I0312 18:18:40.649685 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vj2fp" Mar 12 18:18:41 crc kubenswrapper[4926]: I0312 18:18:41.058162 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-vj2fp"] Mar 12 18:18:41 crc kubenswrapper[4926]: W0312 18:18:41.066534 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7718f63_5b8f_4137_8a80_f8793ce3e60f.slice/crio-d25560d3728b5ebac261cd626d37b593240ae569a7e8134014f5988620070bba WatchSource:0}: Error finding container d25560d3728b5ebac261cd626d37b593240ae569a7e8134014f5988620070bba: Status 404 returned error can't find the container with id d25560d3728b5ebac261cd626d37b593240ae569a7e8134014f5988620070bba Mar 12 18:18:41 crc kubenswrapper[4926]: I0312 18:18:41.130290 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vj2fp" event={"ID":"c7718f63-5b8f-4137-8a80-f8793ce3e60f","Type":"ContainerStarted","Data":"d25560d3728b5ebac261cd626d37b593240ae569a7e8134014f5988620070bba"} Mar 12 18:18:41 crc kubenswrapper[4926]: I0312 18:18:41.134301 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9p72p" event={"ID":"53da3fff-e3f4-4b9d-a887-f5a28f986107","Type":"ContainerStarted","Data":"0df53e49a20875db19a35b9d447319dca5a8e19fed241dcaf1ecbff4e2511284"} Mar 12 18:18:41 crc kubenswrapper[4926]: I0312 18:18:41.134607 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:41 crc kubenswrapper[4926]: I0312 18:18:41.159540 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-9p72p" podStartSLOduration=7.009592831 podStartE2EDuration="16.159526365s" podCreationTimestamp="2026-03-12 18:18:25 +0000 UTC" firstStartedPulling="2026-03-12 18:18:26.780952477 +0000 UTC m=+947.149578810" lastFinishedPulling="2026-03-12 18:18:35.930886001 +0000 UTC m=+956.299512344" observedRunningTime="2026-03-12 18:18:41.15776978 +0000 UTC m=+961.526396143" watchObservedRunningTime="2026-03-12 18:18:41.159526365 +0000 UTC m=+961.528152698" Mar 12 18:18:41 crc kubenswrapper[4926]: I0312 18:18:41.640511 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:41 crc kubenswrapper[4926]: I0312 18:18:41.689212 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:43 crc kubenswrapper[4926]: I0312 18:18:43.667602 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vj2fp"] Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.154993 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vj2fp" event={"ID":"c7718f63-5b8f-4137-8a80-f8793ce3e60f","Type":"ContainerStarted","Data":"78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3"} Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.155484 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-vj2fp" podUID="c7718f63-5b8f-4137-8a80-f8793ce3e60f" containerName="registry-server" containerID="cri-o://78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3" gracePeriod=2 Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.185291 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-vj2fp" podStartSLOduration=1.7441221439999999 podStartE2EDuration="4.185269143s" podCreationTimestamp="2026-03-12 18:18:40 +0000 UTC" firstStartedPulling="2026-03-12 18:18:41.069189505 +0000 UTC m=+961.437815838" lastFinishedPulling="2026-03-12 18:18:43.510336504 +0000 UTC m=+963.878962837" observedRunningTime="2026-03-12 18:18:44.179869314 +0000 UTC m=+964.548495647" watchObservedRunningTime="2026-03-12 18:18:44.185269143 +0000 UTC m=+964.553895506" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.268952 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hwc2c"] Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.270520 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.278763 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hwc2c"] Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.370964 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2h8p\" (UniqueName: \"kubernetes.io/projected/66ae05ba-2fc2-4915-9899-083a49295427-kube-api-access-p2h8p\") pod \"openstack-operator-index-hwc2c\" (UID: \"66ae05ba-2fc2-4915-9899-083a49295427\") " pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.473531 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2h8p\" (UniqueName: \"kubernetes.io/projected/66ae05ba-2fc2-4915-9899-083a49295427-kube-api-access-p2h8p\") pod \"openstack-operator-index-hwc2c\" (UID: \"66ae05ba-2fc2-4915-9899-083a49295427\") " pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.495746 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2h8p\" (UniqueName: \"kubernetes.io/projected/66ae05ba-2fc2-4915-9899-083a49295427-kube-api-access-p2h8p\") pod \"openstack-operator-index-hwc2c\" (UID: \"66ae05ba-2fc2-4915-9899-083a49295427\") " pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.521750 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vj2fp" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.574408 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-js5nb\" (UniqueName: \"kubernetes.io/projected/c7718f63-5b8f-4137-8a80-f8793ce3e60f-kube-api-access-js5nb\") pod \"c7718f63-5b8f-4137-8a80-f8793ce3e60f\" (UID: \"c7718f63-5b8f-4137-8a80-f8793ce3e60f\") " Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.578208 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7718f63-5b8f-4137-8a80-f8793ce3e60f-kube-api-access-js5nb" (OuterVolumeSpecName: "kube-api-access-js5nb") pod "c7718f63-5b8f-4137-8a80-f8793ce3e60f" (UID: "c7718f63-5b8f-4137-8a80-f8793ce3e60f"). InnerVolumeSpecName "kube-api-access-js5nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.621596 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.675711 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-js5nb\" (UniqueName: \"kubernetes.io/projected/c7718f63-5b8f-4137-8a80-f8793ce3e60f-kube-api-access-js5nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:18:44 crc kubenswrapper[4926]: I0312 18:18:44.851791 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hwc2c"] Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.166155 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hwc2c" event={"ID":"66ae05ba-2fc2-4915-9899-083a49295427","Type":"ContainerStarted","Data":"6fc5bf8926e75ff0040ebbba4ef2d374a3641b9d4ffaaa604bc860ce9ed9fd30"} Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.166809 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hwc2c" event={"ID":"66ae05ba-2fc2-4915-9899-083a49295427","Type":"ContainerStarted","Data":"6cccf55ee6fc80f4119c9218e0e40321791ca8a589afdaf0b6a46644e7580f06"} Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.169965 4926 generic.go:334] "Generic (PLEG): container finished" podID="c7718f63-5b8f-4137-8a80-f8793ce3e60f" containerID="78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3" exitCode=0 Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.169994 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-vj2fp" Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.170068 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vj2fp" event={"ID":"c7718f63-5b8f-4137-8a80-f8793ce3e60f","Type":"ContainerDied","Data":"78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3"} Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.170171 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-vj2fp" event={"ID":"c7718f63-5b8f-4137-8a80-f8793ce3e60f","Type":"ContainerDied","Data":"d25560d3728b5ebac261cd626d37b593240ae569a7e8134014f5988620070bba"} Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.170208 4926 scope.go:117] "RemoveContainer" containerID="78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3" Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.192111 4926 scope.go:117] "RemoveContainer" containerID="78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3" Mar 12 18:18:45 crc kubenswrapper[4926]: E0312 18:18:45.193058 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3\": container with ID starting with 78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3 not found: ID does not exist" containerID="78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3" Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.193153 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3"} err="failed to get container status \"78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3\": rpc error: code = NotFound desc = could not find container \"78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3\": container with ID starting with 78cc23ad26fef8004b936ce6e2b86f5bce9f33915b9e8b8f133965ac739d7ff3 not found: ID does not exist" Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.194294 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hwc2c" podStartSLOduration=1.146188808 podStartE2EDuration="1.194270854s" podCreationTimestamp="2026-03-12 18:18:44 +0000 UTC" firstStartedPulling="2026-03-12 18:18:44.877060475 +0000 UTC m=+965.245686808" lastFinishedPulling="2026-03-12 18:18:44.925142521 +0000 UTC m=+965.293768854" observedRunningTime="2026-03-12 18:18:45.18736935 +0000 UTC m=+965.555995723" watchObservedRunningTime="2026-03-12 18:18:45.194270854 +0000 UTC m=+965.562897227" Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.214758 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-vj2fp"] Mar 12 18:18:45 crc kubenswrapper[4926]: I0312 18:18:45.223703 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-vj2fp"] Mar 12 18:18:46 crc kubenswrapper[4926]: I0312 18:18:46.501250 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7718f63-5b8f-4137-8a80-f8793ce3e60f" path="/var/lib/kubelet/pods/c7718f63-5b8f-4137-8a80-f8793ce3e60f/volumes" Mar 12 18:18:46 crc kubenswrapper[4926]: I0312 18:18:46.629224 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-j2n6d" Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.874877 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tn2nt"] Mar 12 18:18:51 crc kubenswrapper[4926]: E0312 18:18:51.875563 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7718f63-5b8f-4137-8a80-f8793ce3e60f" containerName="registry-server" Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.875584 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7718f63-5b8f-4137-8a80-f8793ce3e60f" containerName="registry-server" Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.875829 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7718f63-5b8f-4137-8a80-f8793ce3e60f" containerName="registry-server" Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.877177 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.898023 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tn2nt"] Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.994186 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-catalog-content\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.994254 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6r7m\" (UniqueName: \"kubernetes.io/projected/c3d1b5b0-e990-4051-884a-8b1bfb331618-kube-api-access-h6r7m\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:51 crc kubenswrapper[4926]: I0312 18:18:51.994287 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-utilities\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:52 crc kubenswrapper[4926]: I0312 18:18:52.095061 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-utilities\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:52 crc kubenswrapper[4926]: I0312 18:18:52.095196 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-catalog-content\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:52 crc kubenswrapper[4926]: I0312 18:18:52.095269 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6r7m\" (UniqueName: \"kubernetes.io/projected/c3d1b5b0-e990-4051-884a-8b1bfb331618-kube-api-access-h6r7m\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:52 crc kubenswrapper[4926]: I0312 18:18:52.095827 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-catalog-content\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:52 crc kubenswrapper[4926]: I0312 18:18:52.095849 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-utilities\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:52 crc kubenswrapper[4926]: I0312 18:18:52.132613 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6r7m\" (UniqueName: \"kubernetes.io/projected/c3d1b5b0-e990-4051-884a-8b1bfb331618-kube-api-access-h6r7m\") pod \"community-operators-tn2nt\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:52 crc kubenswrapper[4926]: I0312 18:18:52.226851 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:18:53 crc kubenswrapper[4926]: I0312 18:18:53.192197 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tn2nt"] Mar 12 18:18:53 crc kubenswrapper[4926]: W0312 18:18:53.200690 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3d1b5b0_e990_4051_884a_8b1bfb331618.slice/crio-6f4a9a1bcdab9a6ab28add7cca0c853a121afbf4dfa55c723b2f982597c64d37 WatchSource:0}: Error finding container 6f4a9a1bcdab9a6ab28add7cca0c853a121afbf4dfa55c723b2f982597c64d37: Status 404 returned error can't find the container with id 6f4a9a1bcdab9a6ab28add7cca0c853a121afbf4dfa55c723b2f982597c64d37 Mar 12 18:18:53 crc kubenswrapper[4926]: I0312 18:18:53.249125 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn2nt" event={"ID":"c3d1b5b0-e990-4051-884a-8b1bfb331618","Type":"ContainerStarted","Data":"6f4a9a1bcdab9a6ab28add7cca0c853a121afbf4dfa55c723b2f982597c64d37"} Mar 12 18:18:54 crc kubenswrapper[4926]: I0312 18:18:54.260897 4926 generic.go:334] "Generic (PLEG): container finished" podID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerID="7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88" exitCode=0 Mar 12 18:18:54 crc kubenswrapper[4926]: I0312 18:18:54.262152 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn2nt" event={"ID":"c3d1b5b0-e990-4051-884a-8b1bfb331618","Type":"ContainerDied","Data":"7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88"} Mar 12 18:18:54 crc kubenswrapper[4926]: I0312 18:18:54.622607 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:54 crc kubenswrapper[4926]: I0312 18:18:54.622666 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:54 crc kubenswrapper[4926]: I0312 18:18:54.653657 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:55 crc kubenswrapper[4926]: I0312 18:18:55.315995 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-hwc2c" Mar 12 18:18:56 crc kubenswrapper[4926]: I0312 18:18:56.287811 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn2nt" event={"ID":"c3d1b5b0-e990-4051-884a-8b1bfb331618","Type":"ContainerStarted","Data":"ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3"} Mar 12 18:18:56 crc kubenswrapper[4926]: I0312 18:18:56.642217 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-9p72p" Mar 12 18:18:57 crc kubenswrapper[4926]: I0312 18:18:57.296170 4926 generic.go:334] "Generic (PLEG): container finished" podID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerID="ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3" exitCode=0 Mar 12 18:18:57 crc kubenswrapper[4926]: I0312 18:18:57.296226 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn2nt" event={"ID":"c3d1b5b0-e990-4051-884a-8b1bfb331618","Type":"ContainerDied","Data":"ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3"} Mar 12 18:18:58 crc kubenswrapper[4926]: I0312 18:18:58.306900 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn2nt" event={"ID":"c3d1b5b0-e990-4051-884a-8b1bfb331618","Type":"ContainerStarted","Data":"3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911"} Mar 12 18:19:01 crc kubenswrapper[4926]: I0312 18:19:01.237102 4926 scope.go:117] "RemoveContainer" containerID="80892424cda12377cff2432a54ca031087805d38670f04b3a339fc6f8678512c" Mar 12 18:19:01 crc kubenswrapper[4926]: I0312 18:19:01.936675 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tn2nt" podStartSLOduration=7.484299345 podStartE2EDuration="10.936657235s" podCreationTimestamp="2026-03-12 18:18:51 +0000 UTC" firstStartedPulling="2026-03-12 18:18:54.262818756 +0000 UTC m=+974.631445099" lastFinishedPulling="2026-03-12 18:18:57.715176646 +0000 UTC m=+978.083802989" observedRunningTime="2026-03-12 18:18:58.335065411 +0000 UTC m=+978.703691754" watchObservedRunningTime="2026-03-12 18:19:01.936657235 +0000 UTC m=+982.305283578" Mar 12 18:19:01 crc kubenswrapper[4926]: I0312 18:19:01.939425 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt"] Mar 12 18:19:01 crc kubenswrapper[4926]: I0312 18:19:01.940832 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:01 crc kubenswrapper[4926]: I0312 18:19:01.944271 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-xfbfb" Mar 12 18:19:01 crc kubenswrapper[4926]: I0312 18:19:01.964719 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt"] Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.093779 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-bundle\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.093834 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27czd\" (UniqueName: \"kubernetes.io/projected/1a47463c-c539-4a5c-a3fd-c09feef2de67-kube-api-access-27czd\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.094288 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-util\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.195090 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-util\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.195351 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-bundle\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.195398 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27czd\" (UniqueName: \"kubernetes.io/projected/1a47463c-c539-4a5c-a3fd-c09feef2de67-kube-api-access-27czd\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.196052 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-util\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.196055 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-bundle\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.218279 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27czd\" (UniqueName: \"kubernetes.io/projected/1a47463c-c539-4a5c-a3fd-c09feef2de67-kube-api-access-27czd\") pod \"0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.227397 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.227467 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.263858 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.280737 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.369997 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:19:02 crc kubenswrapper[4926]: I0312 18:19:02.693210 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt"] Mar 12 18:19:02 crc kubenswrapper[4926]: W0312 18:19:02.700664 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a47463c_c539_4a5c_a3fd_c09feef2de67.slice/crio-0a67f29c83a4a856b4c6dea6d1a3b835900776ea48c9acafa40db5354f48f0b3 WatchSource:0}: Error finding container 0a67f29c83a4a856b4c6dea6d1a3b835900776ea48c9acafa40db5354f48f0b3: Status 404 returned error can't find the container with id 0a67f29c83a4a856b4c6dea6d1a3b835900776ea48c9acafa40db5354f48f0b3 Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.273248 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v54tn"] Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.276541 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.295326 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v54tn"] Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.324917 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-utilities\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.324998 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-catalog-content\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.325091 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q26vh\" (UniqueName: \"kubernetes.io/projected/b1bc86da-e204-472f-9744-af11c0e8a265-kube-api-access-q26vh\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.339772 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerID="7db1301999f83d9d375224675ed3da76f79f97b39dcee8b522fb3ee989cf9aa9" exitCode=0 Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.341218 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" event={"ID":"1a47463c-c539-4a5c-a3fd-c09feef2de67","Type":"ContainerDied","Data":"7db1301999f83d9d375224675ed3da76f79f97b39dcee8b522fb3ee989cf9aa9"} Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.341264 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" event={"ID":"1a47463c-c539-4a5c-a3fd-c09feef2de67","Type":"ContainerStarted","Data":"0a67f29c83a4a856b4c6dea6d1a3b835900776ea48c9acafa40db5354f48f0b3"} Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.426057 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-catalog-content\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.426165 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q26vh\" (UniqueName: \"kubernetes.io/projected/b1bc86da-e204-472f-9744-af11c0e8a265-kube-api-access-q26vh\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.426232 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-utilities\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.426834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-utilities\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.427277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-catalog-content\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.455929 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q26vh\" (UniqueName: \"kubernetes.io/projected/b1bc86da-e204-472f-9744-af11c0e8a265-kube-api-access-q26vh\") pod \"certified-operators-v54tn\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.619534 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:03 crc kubenswrapper[4926]: I0312 18:19:03.842446 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v54tn"] Mar 12 18:19:04 crc kubenswrapper[4926]: I0312 18:19:04.350656 4926 generic.go:334] "Generic (PLEG): container finished" podID="b1bc86da-e204-472f-9744-af11c0e8a265" containerID="88e86e609584800aecd00ec27240db80609a2d80782a4ff3069326cd7151ac45" exitCode=0 Mar 12 18:19:04 crc kubenswrapper[4926]: I0312 18:19:04.350735 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v54tn" event={"ID":"b1bc86da-e204-472f-9744-af11c0e8a265","Type":"ContainerDied","Data":"88e86e609584800aecd00ec27240db80609a2d80782a4ff3069326cd7151ac45"} Mar 12 18:19:04 crc kubenswrapper[4926]: I0312 18:19:04.350947 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v54tn" event={"ID":"b1bc86da-e204-472f-9744-af11c0e8a265","Type":"ContainerStarted","Data":"26d978bbd24802db4c024d84e6fb86f4ea3699958fdd595fabcd49191a4e9809"} Mar 12 18:19:04 crc kubenswrapper[4926]: I0312 18:19:04.352667 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerID="ab2124d1cf4401d3e3c56ad6c8da750643ef586b4af6bfe3671baf43b473f054" exitCode=0 Mar 12 18:19:04 crc kubenswrapper[4926]: I0312 18:19:04.352687 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" event={"ID":"1a47463c-c539-4a5c-a3fd-c09feef2de67","Type":"ContainerDied","Data":"ab2124d1cf4401d3e3c56ad6c8da750643ef586b4af6bfe3671baf43b473f054"} Mar 12 18:19:05 crc kubenswrapper[4926]: I0312 18:19:05.359969 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v54tn" event={"ID":"b1bc86da-e204-472f-9744-af11c0e8a265","Type":"ContainerStarted","Data":"8f6bfe79045806d74f84405e9ff1f3a84c03533efa8e6773bb09b52e303ff4e2"} Mar 12 18:19:05 crc kubenswrapper[4926]: I0312 18:19:05.363252 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerID="8da1faac0601d97457d68329ffd4b3a2e386575abb53cf5c7eb97ea42d6b0cd8" exitCode=0 Mar 12 18:19:05 crc kubenswrapper[4926]: I0312 18:19:05.363290 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" event={"ID":"1a47463c-c539-4a5c-a3fd-c09feef2de67","Type":"ContainerDied","Data":"8da1faac0601d97457d68329ffd4b3a2e386575abb53cf5c7eb97ea42d6b0cd8"} Mar 12 18:19:06 crc kubenswrapper[4926]: I0312 18:19:06.375501 4926 generic.go:334] "Generic (PLEG): container finished" podID="b1bc86da-e204-472f-9744-af11c0e8a265" containerID="8f6bfe79045806d74f84405e9ff1f3a84c03533efa8e6773bb09b52e303ff4e2" exitCode=0 Mar 12 18:19:06 crc kubenswrapper[4926]: I0312 18:19:06.375576 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v54tn" event={"ID":"b1bc86da-e204-472f-9744-af11c0e8a265","Type":"ContainerDied","Data":"8f6bfe79045806d74f84405e9ff1f3a84c03533efa8e6773bb09b52e303ff4e2"} Mar 12 18:19:06 crc kubenswrapper[4926]: I0312 18:19:06.848939 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tn2nt"] Mar 12 18:19:06 crc kubenswrapper[4926]: I0312 18:19:06.849212 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tn2nt" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="registry-server" containerID="cri-o://3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911" gracePeriod=2 Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.015723 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.131722 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-bundle\") pod \"1a47463c-c539-4a5c-a3fd-c09feef2de67\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.131818 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27czd\" (UniqueName: \"kubernetes.io/projected/1a47463c-c539-4a5c-a3fd-c09feef2de67-kube-api-access-27czd\") pod \"1a47463c-c539-4a5c-a3fd-c09feef2de67\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.131888 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-util\") pod \"1a47463c-c539-4a5c-a3fd-c09feef2de67\" (UID: \"1a47463c-c539-4a5c-a3fd-c09feef2de67\") " Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.132557 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-bundle" (OuterVolumeSpecName: "bundle") pod "1a47463c-c539-4a5c-a3fd-c09feef2de67" (UID: "1a47463c-c539-4a5c-a3fd-c09feef2de67"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.137668 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a47463c-c539-4a5c-a3fd-c09feef2de67-kube-api-access-27czd" (OuterVolumeSpecName: "kube-api-access-27czd") pod "1a47463c-c539-4a5c-a3fd-c09feef2de67" (UID: "1a47463c-c539-4a5c-a3fd-c09feef2de67"). InnerVolumeSpecName "kube-api-access-27czd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.155289 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-util" (OuterVolumeSpecName: "util") pod "1a47463c-c539-4a5c-a3fd-c09feef2de67" (UID: "1a47463c-c539-4a5c-a3fd-c09feef2de67"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.232771 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-util\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.232801 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a47463c-c539-4a5c-a3fd-c09feef2de67-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.232811 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27czd\" (UniqueName: \"kubernetes.io/projected/1a47463c-c539-4a5c-a3fd-c09feef2de67-kube-api-access-27czd\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.234356 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.333920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-catalog-content\") pod \"c3d1b5b0-e990-4051-884a-8b1bfb331618\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.333998 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6r7m\" (UniqueName: \"kubernetes.io/projected/c3d1b5b0-e990-4051-884a-8b1bfb331618-kube-api-access-h6r7m\") pod \"c3d1b5b0-e990-4051-884a-8b1bfb331618\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.334041 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-utilities\") pod \"c3d1b5b0-e990-4051-884a-8b1bfb331618\" (UID: \"c3d1b5b0-e990-4051-884a-8b1bfb331618\") " Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.334892 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-utilities" (OuterVolumeSpecName: "utilities") pod "c3d1b5b0-e990-4051-884a-8b1bfb331618" (UID: "c3d1b5b0-e990-4051-884a-8b1bfb331618"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.335062 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.337141 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3d1b5b0-e990-4051-884a-8b1bfb331618-kube-api-access-h6r7m" (OuterVolumeSpecName: "kube-api-access-h6r7m") pod "c3d1b5b0-e990-4051-884a-8b1bfb331618" (UID: "c3d1b5b0-e990-4051-884a-8b1bfb331618"). InnerVolumeSpecName "kube-api-access-h6r7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.388640 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v54tn" event={"ID":"b1bc86da-e204-472f-9744-af11c0e8a265","Type":"ContainerStarted","Data":"2a931491103b2d031737572ba24dfcf764dd2ec0c44c2ffec92cc6d03c2bd8ba"} Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.391027 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" event={"ID":"1a47463c-c539-4a5c-a3fd-c09feef2de67","Type":"ContainerDied","Data":"0a67f29c83a4a856b4c6dea6d1a3b835900776ea48c9acafa40db5354f48f0b3"} Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.391059 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a67f29c83a4a856b4c6dea6d1a3b835900776ea48c9acafa40db5354f48f0b3" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.391108 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.395381 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3d1b5b0-e990-4051-884a-8b1bfb331618" (UID: "c3d1b5b0-e990-4051-884a-8b1bfb331618"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.397049 4926 generic.go:334] "Generic (PLEG): container finished" podID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerID="3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911" exitCode=0 Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.397079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn2nt" event={"ID":"c3d1b5b0-e990-4051-884a-8b1bfb331618","Type":"ContainerDied","Data":"3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911"} Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.397099 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tn2nt" event={"ID":"c3d1b5b0-e990-4051-884a-8b1bfb331618","Type":"ContainerDied","Data":"6f4a9a1bcdab9a6ab28add7cca0c853a121afbf4dfa55c723b2f982597c64d37"} Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.397115 4926 scope.go:117] "RemoveContainer" containerID="3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.397213 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tn2nt" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.428130 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v54tn" podStartSLOduration=1.8928752690000001 podStartE2EDuration="4.428113045s" podCreationTimestamp="2026-03-12 18:19:03 +0000 UTC" firstStartedPulling="2026-03-12 18:19:04.352776585 +0000 UTC m=+984.721402918" lastFinishedPulling="2026-03-12 18:19:06.888014321 +0000 UTC m=+987.256640694" observedRunningTime="2026-03-12 18:19:07.424239605 +0000 UTC m=+987.792865938" watchObservedRunningTime="2026-03-12 18:19:07.428113045 +0000 UTC m=+987.796739378" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.431812 4926 scope.go:117] "RemoveContainer" containerID="ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.444845 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6r7m\" (UniqueName: \"kubernetes.io/projected/c3d1b5b0-e990-4051-884a-8b1bfb331618-kube-api-access-h6r7m\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.444983 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3d1b5b0-e990-4051-884a-8b1bfb331618-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.447471 4926 scope.go:117] "RemoveContainer" containerID="7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.464951 4926 scope.go:117] "RemoveContainer" containerID="3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911" Mar 12 18:19:07 crc kubenswrapper[4926]: E0312 18:19:07.465508 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911\": container with ID starting with 3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911 not found: ID does not exist" containerID="3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.465537 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911"} err="failed to get container status \"3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911\": rpc error: code = NotFound desc = could not find container \"3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911\": container with ID starting with 3c291b723d1fcde95eb414f900dad95e6d41bc5c05e6a755da9ff76633773911 not found: ID does not exist" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.465560 4926 scope.go:117] "RemoveContainer" containerID="ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.465622 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tn2nt"] Mar 12 18:19:07 crc kubenswrapper[4926]: E0312 18:19:07.465763 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3\": container with ID starting with ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3 not found: ID does not exist" containerID="ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.465792 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3"} err="failed to get container status \"ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3\": rpc error: code = NotFound desc = could not find container \"ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3\": container with ID starting with ab4205b30f9b012dcfa5b4992c8916cbd40cecadcc8d510b5930cd515376aff3 not found: ID does not exist" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.465828 4926 scope.go:117] "RemoveContainer" containerID="7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88" Mar 12 18:19:07 crc kubenswrapper[4926]: E0312 18:19:07.466627 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88\": container with ID starting with 7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88 not found: ID does not exist" containerID="7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.466666 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88"} err="failed to get container status \"7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88\": rpc error: code = NotFound desc = could not find container \"7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88\": container with ID starting with 7379ac14c87898826abda5e79ef7d089396a096b81ae86a0cd8ad5065ea31f88 not found: ID does not exist" Mar 12 18:19:07 crc kubenswrapper[4926]: I0312 18:19:07.471483 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tn2nt"] Mar 12 18:19:08 crc kubenswrapper[4926]: I0312 18:19:08.500474 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" path="/var/lib/kubelet/pods/c3d1b5b0-e990-4051-884a-8b1bfb331618/volumes" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.056517 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-66f4595798-c276m"] Mar 12 18:19:13 crc kubenswrapper[4926]: E0312 18:19:13.057101 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="extract-utilities" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057115 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="extract-utilities" Mar 12 18:19:13 crc kubenswrapper[4926]: E0312 18:19:13.057127 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerName="util" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057134 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerName="util" Mar 12 18:19:13 crc kubenswrapper[4926]: E0312 18:19:13.057142 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerName="pull" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057150 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerName="pull" Mar 12 18:19:13 crc kubenswrapper[4926]: E0312 18:19:13.057166 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="extract-content" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057173 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="extract-content" Mar 12 18:19:13 crc kubenswrapper[4926]: E0312 18:19:13.057184 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="registry-server" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057191 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="registry-server" Mar 12 18:19:13 crc kubenswrapper[4926]: E0312 18:19:13.057209 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerName="extract" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057216 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerName="extract" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057342 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3d1b5b0-e990-4051-884a-8b1bfb331618" containerName="registry-server" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057360 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a47463c-c539-4a5c-a3fd-c09feef2de67" containerName="extract" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.057870 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.060193 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-q8skp" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.080864 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-66f4595798-c276m"] Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.117476 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km57g\" (UniqueName: \"kubernetes.io/projected/fe4ee666-a3c1-44c5-a07b-5ca8438e0482-kube-api-access-km57g\") pod \"openstack-operator-controller-init-66f4595798-c276m\" (UID: \"fe4ee666-a3c1-44c5-a07b-5ca8438e0482\") " pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.218818 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km57g\" (UniqueName: \"kubernetes.io/projected/fe4ee666-a3c1-44c5-a07b-5ca8438e0482-kube-api-access-km57g\") pod \"openstack-operator-controller-init-66f4595798-c276m\" (UID: \"fe4ee666-a3c1-44c5-a07b-5ca8438e0482\") " pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.240177 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km57g\" (UniqueName: \"kubernetes.io/projected/fe4ee666-a3c1-44c5-a07b-5ca8438e0482-kube-api-access-km57g\") pod \"openstack-operator-controller-init-66f4595798-c276m\" (UID: \"fe4ee666-a3c1-44c5-a07b-5ca8438e0482\") " pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.378186 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.615173 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-66f4595798-c276m"] Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.620761 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.620989 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.668971 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.874241 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lkpwt"] Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.877259 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.882757 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lkpwt"] Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.930823 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp7sb\" (UniqueName: \"kubernetes.io/projected/5760df4b-1304-4183-8ab8-9fbfb9bf834e-kube-api-access-wp7sb\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.930955 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-catalog-content\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:13 crc kubenswrapper[4926]: I0312 18:19:13.931002 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-utilities\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.031526 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-catalog-content\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.031572 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-utilities\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.031627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp7sb\" (UniqueName: \"kubernetes.io/projected/5760df4b-1304-4183-8ab8-9fbfb9bf834e-kube-api-access-wp7sb\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.032285 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-utilities\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.032312 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-catalog-content\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.051145 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp7sb\" (UniqueName: \"kubernetes.io/projected/5760df4b-1304-4183-8ab8-9fbfb9bf834e-kube-api-access-wp7sb\") pod \"redhat-marketplace-lkpwt\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.202466 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.398912 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lkpwt"] Mar 12 18:19:14 crc kubenswrapper[4926]: W0312 18:19:14.414855 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5760df4b_1304_4183_8ab8_9fbfb9bf834e.slice/crio-e12189e32000defc9737b3f0fdd65ebb428033f60ea171fd2cf14de8505756de WatchSource:0}: Error finding container e12189e32000defc9737b3f0fdd65ebb428033f60ea171fd2cf14de8505756de: Status 404 returned error can't find the container with id e12189e32000defc9737b3f0fdd65ebb428033f60ea171fd2cf14de8505756de Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.448111 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lkpwt" event={"ID":"5760df4b-1304-4183-8ab8-9fbfb9bf834e","Type":"ContainerStarted","Data":"e12189e32000defc9737b3f0fdd65ebb428033f60ea171fd2cf14de8505756de"} Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.449665 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" event={"ID":"fe4ee666-a3c1-44c5-a07b-5ca8438e0482","Type":"ContainerStarted","Data":"d532efc1ea779c2d60ac6c2534240645c07ed3d7a7cd4a1d3ecae9187aefb3b9"} Mar 12 18:19:14 crc kubenswrapper[4926]: I0312 18:19:14.502225 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:15 crc kubenswrapper[4926]: I0312 18:19:15.456980 4926 generic.go:334] "Generic (PLEG): container finished" podID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerID="ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb" exitCode=0 Mar 12 18:19:15 crc kubenswrapper[4926]: I0312 18:19:15.457081 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lkpwt" event={"ID":"5760df4b-1304-4183-8ab8-9fbfb9bf834e","Type":"ContainerDied","Data":"ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb"} Mar 12 18:19:16 crc kubenswrapper[4926]: I0312 18:19:16.449140 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v54tn"] Mar 12 18:19:17 crc kubenswrapper[4926]: I0312 18:19:17.482089 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v54tn" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="registry-server" containerID="cri-o://2a931491103b2d031737572ba24dfcf764dd2ec0c44c2ffec92cc6d03c2bd8ba" gracePeriod=2 Mar 12 18:19:18 crc kubenswrapper[4926]: I0312 18:19:18.491053 4926 generic.go:334] "Generic (PLEG): container finished" podID="b1bc86da-e204-472f-9744-af11c0e8a265" containerID="2a931491103b2d031737572ba24dfcf764dd2ec0c44c2ffec92cc6d03c2bd8ba" exitCode=0 Mar 12 18:19:18 crc kubenswrapper[4926]: I0312 18:19:18.500886 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v54tn" event={"ID":"b1bc86da-e204-472f-9744-af11c0e8a265","Type":"ContainerDied","Data":"2a931491103b2d031737572ba24dfcf764dd2ec0c44c2ffec92cc6d03c2bd8ba"} Mar 12 18:19:19 crc kubenswrapper[4926]: I0312 18:19:19.905553 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.039799 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-catalog-content\") pod \"b1bc86da-e204-472f-9744-af11c0e8a265\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.040128 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-utilities\") pod \"b1bc86da-e204-472f-9744-af11c0e8a265\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.040274 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q26vh\" (UniqueName: \"kubernetes.io/projected/b1bc86da-e204-472f-9744-af11c0e8a265-kube-api-access-q26vh\") pod \"b1bc86da-e204-472f-9744-af11c0e8a265\" (UID: \"b1bc86da-e204-472f-9744-af11c0e8a265\") " Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.043499 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-utilities" (OuterVolumeSpecName: "utilities") pod "b1bc86da-e204-472f-9744-af11c0e8a265" (UID: "b1bc86da-e204-472f-9744-af11c0e8a265"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.047137 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1bc86da-e204-472f-9744-af11c0e8a265-kube-api-access-q26vh" (OuterVolumeSpecName: "kube-api-access-q26vh") pod "b1bc86da-e204-472f-9744-af11c0e8a265" (UID: "b1bc86da-e204-472f-9744-af11c0e8a265"). InnerVolumeSpecName "kube-api-access-q26vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.098270 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1bc86da-e204-472f-9744-af11c0e8a265" (UID: "b1bc86da-e204-472f-9744-af11c0e8a265"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.141916 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.141955 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1bc86da-e204-472f-9744-af11c0e8a265-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.141983 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q26vh\" (UniqueName: \"kubernetes.io/projected/b1bc86da-e204-472f-9744-af11c0e8a265-kube-api-access-q26vh\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.505210 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v54tn" event={"ID":"b1bc86da-e204-472f-9744-af11c0e8a265","Type":"ContainerDied","Data":"26d978bbd24802db4c024d84e6fb86f4ea3699958fdd595fabcd49191a4e9809"} Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.505284 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v54tn" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.505301 4926 scope.go:117] "RemoveContainer" containerID="2a931491103b2d031737572ba24dfcf764dd2ec0c44c2ffec92cc6d03c2bd8ba" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.507143 4926 generic.go:334] "Generic (PLEG): container finished" podID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerID="321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d" exitCode=0 Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.507246 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lkpwt" event={"ID":"5760df4b-1304-4183-8ab8-9fbfb9bf834e","Type":"ContainerDied","Data":"321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d"} Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.510957 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" event={"ID":"fe4ee666-a3c1-44c5-a07b-5ca8438e0482","Type":"ContainerStarted","Data":"919f28d38b2ea6bc9cc52444fe8a3644659066c1d1bad126cd7ef0693ae5e8f1"} Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.512082 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.530327 4926 scope.go:117] "RemoveContainer" containerID="8f6bfe79045806d74f84405e9ff1f3a84c03533efa8e6773bb09b52e303ff4e2" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.543822 4926 scope.go:117] "RemoveContainer" containerID="88e86e609584800aecd00ec27240db80609a2d80782a4ff3069326cd7151ac45" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.563695 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" podStartSLOduration=1.285328016 podStartE2EDuration="7.563670999s" podCreationTimestamp="2026-03-12 18:19:13 +0000 UTC" firstStartedPulling="2026-03-12 18:19:13.626251352 +0000 UTC m=+993.994877685" lastFinishedPulling="2026-03-12 18:19:19.904594335 +0000 UTC m=+1000.273220668" observedRunningTime="2026-03-12 18:19:20.559143739 +0000 UTC m=+1000.927770072" watchObservedRunningTime="2026-03-12 18:19:20.563670999 +0000 UTC m=+1000.932297372" Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.590486 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v54tn"] Mar 12 18:19:20 crc kubenswrapper[4926]: I0312 18:19:20.594219 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v54tn"] Mar 12 18:19:21 crc kubenswrapper[4926]: I0312 18:19:21.521116 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lkpwt" event={"ID":"5760df4b-1304-4183-8ab8-9fbfb9bf834e","Type":"ContainerStarted","Data":"8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171"} Mar 12 18:19:21 crc kubenswrapper[4926]: I0312 18:19:21.541956 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lkpwt" podStartSLOduration=3.507176673 podStartE2EDuration="8.541938166s" podCreationTimestamp="2026-03-12 18:19:13 +0000 UTC" firstStartedPulling="2026-03-12 18:19:16.098075235 +0000 UTC m=+996.466701568" lastFinishedPulling="2026-03-12 18:19:21.132836728 +0000 UTC m=+1001.501463061" observedRunningTime="2026-03-12 18:19:21.537359593 +0000 UTC m=+1001.905985916" watchObservedRunningTime="2026-03-12 18:19:21.541938166 +0000 UTC m=+1001.910564499" Mar 12 18:19:22 crc kubenswrapper[4926]: I0312 18:19:22.497653 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" path="/var/lib/kubelet/pods/b1bc86da-e204-472f-9744-af11c0e8a265/volumes" Mar 12 18:19:24 crc kubenswrapper[4926]: I0312 18:19:24.203290 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:24 crc kubenswrapper[4926]: I0312 18:19:24.204474 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:24 crc kubenswrapper[4926]: I0312 18:19:24.279814 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:33 crc kubenswrapper[4926]: I0312 18:19:33.382085 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-66f4595798-c276m" Mar 12 18:19:34 crc kubenswrapper[4926]: I0312 18:19:34.298077 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:34 crc kubenswrapper[4926]: I0312 18:19:34.338294 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lkpwt"] Mar 12 18:19:34 crc kubenswrapper[4926]: I0312 18:19:34.612218 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lkpwt" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="registry-server" containerID="cri-o://8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171" gracePeriod=2 Mar 12 18:19:34 crc kubenswrapper[4926]: I0312 18:19:34.974884 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.090535 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-catalog-content\") pod \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.092707 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-utilities\") pod \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.092770 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp7sb\" (UniqueName: \"kubernetes.io/projected/5760df4b-1304-4183-8ab8-9fbfb9bf834e-kube-api-access-wp7sb\") pod \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\" (UID: \"5760df4b-1304-4183-8ab8-9fbfb9bf834e\") " Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.094549 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-utilities" (OuterVolumeSpecName: "utilities") pod "5760df4b-1304-4183-8ab8-9fbfb9bf834e" (UID: "5760df4b-1304-4183-8ab8-9fbfb9bf834e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.099268 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5760df4b-1304-4183-8ab8-9fbfb9bf834e-kube-api-access-wp7sb" (OuterVolumeSpecName: "kube-api-access-wp7sb") pod "5760df4b-1304-4183-8ab8-9fbfb9bf834e" (UID: "5760df4b-1304-4183-8ab8-9fbfb9bf834e"). InnerVolumeSpecName "kube-api-access-wp7sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.126288 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5760df4b-1304-4183-8ab8-9fbfb9bf834e" (UID: "5760df4b-1304-4183-8ab8-9fbfb9bf834e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.195194 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.195258 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp7sb\" (UniqueName: \"kubernetes.io/projected/5760df4b-1304-4183-8ab8-9fbfb9bf834e-kube-api-access-wp7sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.195282 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5760df4b-1304-4183-8ab8-9fbfb9bf834e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.623761 4926 generic.go:334] "Generic (PLEG): container finished" podID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerID="8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171" exitCode=0 Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.623834 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lkpwt" event={"ID":"5760df4b-1304-4183-8ab8-9fbfb9bf834e","Type":"ContainerDied","Data":"8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171"} Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.623861 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lkpwt" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.623898 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lkpwt" event={"ID":"5760df4b-1304-4183-8ab8-9fbfb9bf834e","Type":"ContainerDied","Data":"e12189e32000defc9737b3f0fdd65ebb428033f60ea171fd2cf14de8505756de"} Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.623939 4926 scope.go:117] "RemoveContainer" containerID="8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.658744 4926 scope.go:117] "RemoveContainer" containerID="321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.662840 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lkpwt"] Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.670317 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lkpwt"] Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.695819 4926 scope.go:117] "RemoveContainer" containerID="ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.714523 4926 scope.go:117] "RemoveContainer" containerID="8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171" Mar 12 18:19:35 crc kubenswrapper[4926]: E0312 18:19:35.714851 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171\": container with ID starting with 8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171 not found: ID does not exist" containerID="8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.714904 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171"} err="failed to get container status \"8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171\": rpc error: code = NotFound desc = could not find container \"8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171\": container with ID starting with 8cfe4fe1a1488acebc531e1b4876a26cd083167fab9b1c38ab77343ef7244171 not found: ID does not exist" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.714946 4926 scope.go:117] "RemoveContainer" containerID="321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d" Mar 12 18:19:35 crc kubenswrapper[4926]: E0312 18:19:35.715323 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d\": container with ID starting with 321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d not found: ID does not exist" containerID="321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.715363 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d"} err="failed to get container status \"321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d\": rpc error: code = NotFound desc = could not find container \"321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d\": container with ID starting with 321acb9fc46bde9a355bab5d1b06a1220bab5d4dee54e97656cf70a21d113f6d not found: ID does not exist" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.715385 4926 scope.go:117] "RemoveContainer" containerID="ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb" Mar 12 18:19:35 crc kubenswrapper[4926]: E0312 18:19:35.715660 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb\": container with ID starting with ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb not found: ID does not exist" containerID="ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb" Mar 12 18:19:35 crc kubenswrapper[4926]: I0312 18:19:35.715689 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb"} err="failed to get container status \"ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb\": rpc error: code = NotFound desc = could not find container \"ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb\": container with ID starting with ed4aa3e456adb49e71017dd6c15eb8257568bdc74ddf99e569ac60e8d26b5ffb not found: ID does not exist" Mar 12 18:19:36 crc kubenswrapper[4926]: I0312 18:19:36.502697 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" path="/var/lib/kubelet/pods/5760df4b-1304-4183-8ab8-9fbfb9bf834e/volumes" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.134023 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555660-cbx5b"] Mar 12 18:20:00 crc kubenswrapper[4926]: E0312 18:20:00.135005 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="registry-server" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135024 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="registry-server" Mar 12 18:20:00 crc kubenswrapper[4926]: E0312 18:20:00.135043 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="extract-content" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135053 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="extract-content" Mar 12 18:20:00 crc kubenswrapper[4926]: E0312 18:20:00.135070 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="registry-server" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135081 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="registry-server" Mar 12 18:20:00 crc kubenswrapper[4926]: E0312 18:20:00.135121 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="extract-utilities" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135132 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="extract-utilities" Mar 12 18:20:00 crc kubenswrapper[4926]: E0312 18:20:00.135147 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="extract-content" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135159 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="extract-content" Mar 12 18:20:00 crc kubenswrapper[4926]: E0312 18:20:00.135178 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="extract-utilities" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135188 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="extract-utilities" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135355 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1bc86da-e204-472f-9744-af11c0e8a265" containerName="registry-server" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135371 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5760df4b-1304-4183-8ab8-9fbfb9bf834e" containerName="registry-server" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.135990 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555660-cbx5b" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.138293 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.138670 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.139180 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.145465 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555660-cbx5b"] Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.268455 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6s5b\" (UniqueName: \"kubernetes.io/projected/b50fb579-57d6-4029-a4f3-c8a3303bac4d-kube-api-access-f6s5b\") pod \"auto-csr-approver-29555660-cbx5b\" (UID: \"b50fb579-57d6-4029-a4f3-c8a3303bac4d\") " pod="openshift-infra/auto-csr-approver-29555660-cbx5b" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.370334 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6s5b\" (UniqueName: \"kubernetes.io/projected/b50fb579-57d6-4029-a4f3-c8a3303bac4d-kube-api-access-f6s5b\") pod \"auto-csr-approver-29555660-cbx5b\" (UID: \"b50fb579-57d6-4029-a4f3-c8a3303bac4d\") " pod="openshift-infra/auto-csr-approver-29555660-cbx5b" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.396271 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6s5b\" (UniqueName: \"kubernetes.io/projected/b50fb579-57d6-4029-a4f3-c8a3303bac4d-kube-api-access-f6s5b\") pod \"auto-csr-approver-29555660-cbx5b\" (UID: \"b50fb579-57d6-4029-a4f3-c8a3303bac4d\") " pod="openshift-infra/auto-csr-approver-29555660-cbx5b" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.453958 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555660-cbx5b" Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.687580 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555660-cbx5b"] Mar 12 18:20:00 crc kubenswrapper[4926]: I0312 18:20:00.788232 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555660-cbx5b" event={"ID":"b50fb579-57d6-4029-a4f3-c8a3303bac4d","Type":"ContainerStarted","Data":"f263435b763cc115efbb77597e4b93b8d2c4a1f2f6ba06fd3f2f757eb1164e16"} Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.436323 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.437759 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.445935 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-5lgf2" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.449162 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.450172 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.453666 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.458008 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.458330 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-9x7p7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.458720 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.460119 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-j6zqr" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.476337 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.480924 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.486655 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.487554 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.491650 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-nwrjb" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.498678 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phfsx\" (UniqueName: \"kubernetes.io/projected/e7b3fba0-ddaa-4cef-9df6-0683a92475cf-kube-api-access-phfsx\") pod \"barbican-operator-controller-manager-677bd678f7-ssmx7\" (UID: \"e7b3fba0-ddaa-4cef-9df6-0683a92475cf\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.498771 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85cqv\" (UniqueName: \"kubernetes.io/projected/f112cb87-7454-41fa-a1e1-381d79f86247-kube-api-access-85cqv\") pod \"cinder-operator-controller-manager-984cd4dcf-zlsrd\" (UID: \"f112cb87-7454-41fa-a1e1-381d79f86247\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.509213 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-525z5"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.514080 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.517920 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-xnnxs" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.546013 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.575772 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-525z5"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.587432 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.588357 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.591188 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-tdvjg" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.599693 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2zjq\" (UniqueName: \"kubernetes.io/projected/4800992f-cfad-4a1c-94e5-79427f88c002-kube-api-access-n2zjq\") pod \"designate-operator-controller-manager-66d56f6ff4-clngf\" (UID: \"4800992f-cfad-4a1c-94e5-79427f88c002\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.599751 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phfsx\" (UniqueName: \"kubernetes.io/projected/e7b3fba0-ddaa-4cef-9df6-0683a92475cf-kube-api-access-phfsx\") pod \"barbican-operator-controller-manager-677bd678f7-ssmx7\" (UID: \"e7b3fba0-ddaa-4cef-9df6-0683a92475cf\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.599790 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d55vh\" (UniqueName: \"kubernetes.io/projected/fea71415-42ac-4e77-ba9c-25170ccece27-kube-api-access-d55vh\") pod \"heat-operator-controller-manager-77b6666d85-525z5\" (UID: \"fea71415-42ac-4e77-ba9c-25170ccece27\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.599904 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p26t\" (UniqueName: \"kubernetes.io/projected/b776be98-1352-43c6-8ee8-e31076b7d12b-kube-api-access-7p26t\") pod \"glance-operator-controller-manager-5964f64c48-m6sxh\" (UID: \"b776be98-1352-43c6-8ee8-e31076b7d12b\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.599934 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85cqv\" (UniqueName: \"kubernetes.io/projected/f112cb87-7454-41fa-a1e1-381d79f86247-kube-api-access-85cqv\") pod \"cinder-operator-controller-manager-984cd4dcf-zlsrd\" (UID: \"f112cb87-7454-41fa-a1e1-381d79f86247\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.634620 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.641806 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85cqv\" (UniqueName: \"kubernetes.io/projected/f112cb87-7454-41fa-a1e1-381d79f86247-kube-api-access-85cqv\") pod \"cinder-operator-controller-manager-984cd4dcf-zlsrd\" (UID: \"f112cb87-7454-41fa-a1e1-381d79f86247\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.642520 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phfsx\" (UniqueName: \"kubernetes.io/projected/e7b3fba0-ddaa-4cef-9df6-0683a92475cf-kube-api-access-phfsx\") pod \"barbican-operator-controller-manager-677bd678f7-ssmx7\" (UID: \"e7b3fba0-ddaa-4cef-9df6-0683a92475cf\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.648592 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.649542 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.653733 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.655415 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-mqcdw" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.673755 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.674719 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.677210 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-rbkds" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.681604 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.700931 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.701676 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p26t\" (UniqueName: \"kubernetes.io/projected/b776be98-1352-43c6-8ee8-e31076b7d12b-kube-api-access-7p26t\") pod \"glance-operator-controller-manager-5964f64c48-m6sxh\" (UID: \"b776be98-1352-43c6-8ee8-e31076b7d12b\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.701726 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg25f\" (UniqueName: \"kubernetes.io/projected/065fe73a-651c-4cd3-b8d7-135617c51bbd-kube-api-access-fg25f\") pod \"horizon-operator-controller-manager-6d9d6b584d-p76mx\" (UID: \"065fe73a-651c-4cd3-b8d7-135617c51bbd\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.701762 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2zjq\" (UniqueName: \"kubernetes.io/projected/4800992f-cfad-4a1c-94e5-79427f88c002-kube-api-access-n2zjq\") pod \"designate-operator-controller-manager-66d56f6ff4-clngf\" (UID: \"4800992f-cfad-4a1c-94e5-79427f88c002\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.701791 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d55vh\" (UniqueName: \"kubernetes.io/projected/fea71415-42ac-4e77-ba9c-25170ccece27-kube-api-access-d55vh\") pod \"heat-operator-controller-manager-77b6666d85-525z5\" (UID: \"fea71415-42ac-4e77-ba9c-25170ccece27\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.701816 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.701834 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwkzx\" (UniqueName: \"kubernetes.io/projected/fd35525c-7b73-49d1-a36c-c49d3bf933eb-kube-api-access-hwkzx\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.705048 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.715376 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.716692 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.723587 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hqcpj" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.723587 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.724739 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.740025 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p26t\" (UniqueName: \"kubernetes.io/projected/b776be98-1352-43c6-8ee8-e31076b7d12b-kube-api-access-7p26t\") pod \"glance-operator-controller-manager-5964f64c48-m6sxh\" (UID: \"b776be98-1352-43c6-8ee8-e31076b7d12b\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.740398 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-cf7zh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.751056 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d55vh\" (UniqueName: \"kubernetes.io/projected/fea71415-42ac-4e77-ba9c-25170ccece27-kube-api-access-d55vh\") pod \"heat-operator-controller-manager-77b6666d85-525z5\" (UID: \"fea71415-42ac-4e77-ba9c-25170ccece27\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.753994 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2zjq\" (UniqueName: \"kubernetes.io/projected/4800992f-cfad-4a1c-94e5-79427f88c002-kube-api-access-n2zjq\") pod \"designate-operator-controller-manager-66d56f6ff4-clngf\" (UID: \"4800992f-cfad-4a1c-94e5-79427f88c002\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.764600 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.764941 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.784732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.786511 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.787603 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.791731 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-bphvw" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.808789 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.808835 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwkzx\" (UniqueName: \"kubernetes.io/projected/fd35525c-7b73-49d1-a36c-c49d3bf933eb-kube-api-access-hwkzx\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.808879 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pcf6\" (UniqueName: \"kubernetes.io/projected/c62baaa0-d72b-4240-9dba-858bdf61d1b3-kube-api-access-4pcf6\") pod \"keystone-operator-controller-manager-684f77d66d-5f6c4\" (UID: \"c62baaa0-d72b-4240-9dba-858bdf61d1b3\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.808895 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" Mar 12 18:20:02 crc kubenswrapper[4926]: E0312 18:20:02.809014 4926 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:02 crc kubenswrapper[4926]: E0312 18:20:02.809058 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert podName:fd35525c-7b73-49d1-a36c-c49d3bf933eb nodeName:}" failed. No retries permitted until 2026-03-12 18:20:03.30904156 +0000 UTC m=+1043.677667893 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert") pod "infra-operator-controller-manager-5995f4446f-kzl7p" (UID: "fd35525c-7b73-49d1-a36c-c49d3bf933eb") : secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.808907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9sjn\" (UniqueName: \"kubernetes.io/projected/a41da562-a119-4785-95d0-eaf0970a99f4-kube-api-access-d9sjn\") pod \"ironic-operator-controller-manager-6bbb499bbc-7q8xr\" (UID: \"a41da562-a119-4785-95d0-eaf0970a99f4\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.809315 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg25f\" (UniqueName: \"kubernetes.io/projected/065fe73a-651c-4cd3-b8d7-135617c51bbd-kube-api-access-fg25f\") pod \"horizon-operator-controller-manager-6d9d6b584d-p76mx\" (UID: \"065fe73a-651c-4cd3-b8d7-135617c51bbd\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.809374 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4g6t\" (UniqueName: \"kubernetes.io/projected/e38cf931-bbd6-4b47-bdf6-8a514d17d3d7-kube-api-access-n4g6t\") pod \"manila-operator-controller-manager-68f45f9d9f-f96w7\" (UID: \"e38cf931-bbd6-4b47-bdf6-8a514d17d3d7\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.831647 4926 generic.go:334] "Generic (PLEG): container finished" podID="b50fb579-57d6-4029-a4f3-c8a3303bac4d" containerID="6f2e2b241a59eb4ec9fee306fab9cd670167a7fcd068ac1d7f97bc699a2c0ee6" exitCode=0 Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.831695 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555660-cbx5b" event={"ID":"b50fb579-57d6-4029-a4f3-c8a3303bac4d","Type":"ContainerDied","Data":"6f2e2b241a59eb4ec9fee306fab9cd670167a7fcd068ac1d7f97bc699a2c0ee6"} Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.844227 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.846596 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.859362 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.860131 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.860519 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.866300 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-wdqjw" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.868309 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.869234 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.872940 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.876963 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-x9sdm" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.883266 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.887395 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg25f\" (UniqueName: \"kubernetes.io/projected/065fe73a-651c-4cd3-b8d7-135617c51bbd-kube-api-access-fg25f\") pod \"horizon-operator-controller-manager-6d9d6b584d-p76mx\" (UID: \"065fe73a-651c-4cd3-b8d7-135617c51bbd\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.888370 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-2fqgq" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.893079 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwkzx\" (UniqueName: \"kubernetes.io/projected/fd35525c-7b73-49d1-a36c-c49d3bf933eb-kube-api-access-hwkzx\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.894615 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.904426 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.908902 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.913342 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.914282 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5wbk\" (UniqueName: \"kubernetes.io/projected/2e44c177-b87d-4ff6-80ca-672477fe9e94-kube-api-access-q5wbk\") pod \"nova-operator-controller-manager-569cc54c5-dzj5w\" (UID: \"2e44c177-b87d-4ff6-80ca-672477fe9e94\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.914324 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q249w\" (UniqueName: \"kubernetes.io/projected/ccd465e8-6811-4865-9602-3dea8144cc01-kube-api-access-q249w\") pod \"mariadb-operator-controller-manager-658d4cdd5-c58lh\" (UID: \"ccd465e8-6811-4865-9602-3dea8144cc01\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.914352 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4g6t\" (UniqueName: \"kubernetes.io/projected/e38cf931-bbd6-4b47-bdf6-8a514d17d3d7-kube-api-access-n4g6t\") pod \"manila-operator-controller-manager-68f45f9d9f-f96w7\" (UID: \"e38cf931-bbd6-4b47-bdf6-8a514d17d3d7\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.914481 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rslhc\" (UniqueName: \"kubernetes.io/projected/6374b396-00ef-4aca-ac07-fd46982f23f1-kube-api-access-rslhc\") pod \"neutron-operator-controller-manager-776c5696bf-njrjb\" (UID: \"6374b396-00ef-4aca-ac07-fd46982f23f1\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.914529 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pcf6\" (UniqueName: \"kubernetes.io/projected/c62baaa0-d72b-4240-9dba-858bdf61d1b3-kube-api-access-4pcf6\") pod \"keystone-operator-controller-manager-684f77d66d-5f6c4\" (UID: \"c62baaa0-d72b-4240-9dba-858bdf61d1b3\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.914568 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9sjn\" (UniqueName: \"kubernetes.io/projected/a41da562-a119-4785-95d0-eaf0970a99f4-kube-api-access-d9sjn\") pod \"ironic-operator-controller-manager-6bbb499bbc-7q8xr\" (UID: \"a41da562-a119-4785-95d0-eaf0970a99f4\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.917494 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.918602 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.921368 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-drhd2" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.930507 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.931546 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.935933 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-49m2z" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.940258 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.941204 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.955243 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.956685 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.956819 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.959602 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-5jlln" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.959792 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.963214 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.966691 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4g6t\" (UniqueName: \"kubernetes.io/projected/e38cf931-bbd6-4b47-bdf6-8a514d17d3d7-kube-api-access-n4g6t\") pod \"manila-operator-controller-manager-68f45f9d9f-f96w7\" (UID: \"e38cf931-bbd6-4b47-bdf6-8a514d17d3d7\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.969740 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9sjn\" (UniqueName: \"kubernetes.io/projected/a41da562-a119-4785-95d0-eaf0970a99f4-kube-api-access-d9sjn\") pod \"ironic-operator-controller-manager-6bbb499bbc-7q8xr\" (UID: \"a41da562-a119-4785-95d0-eaf0970a99f4\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.970762 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747"] Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.979781 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-gglcr" Mar 12 18:20:02 crc kubenswrapper[4926]: I0312 18:20:02.985292 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pcf6\" (UniqueName: \"kubernetes.io/projected/c62baaa0-d72b-4240-9dba-858bdf61d1b3-kube-api-access-4pcf6\") pod \"keystone-operator-controller-manager-684f77d66d-5f6c4\" (UID: \"c62baaa0-d72b-4240-9dba-858bdf61d1b3\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.002760 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.032517 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f48c2\" (UniqueName: \"kubernetes.io/projected/14346445-95be-488f-858c-44bf5b45c656-kube-api-access-f48c2\") pod \"ovn-operator-controller-manager-bbc5b68f9-xr747\" (UID: \"14346445-95be-488f-858c-44bf5b45c656\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.032613 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg8wk\" (UniqueName: \"kubernetes.io/projected/e3baa344-9dd8-48ed-8b6a-60ff9fbc181a-kube-api-access-hg8wk\") pod \"placement-operator-controller-manager-574d45c66c-tpnlt\" (UID: \"e3baa344-9dd8-48ed-8b6a-60ff9fbc181a\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.032673 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rslhc\" (UniqueName: \"kubernetes.io/projected/6374b396-00ef-4aca-ac07-fd46982f23f1-kube-api-access-rslhc\") pod \"neutron-operator-controller-manager-776c5696bf-njrjb\" (UID: \"6374b396-00ef-4aca-ac07-fd46982f23f1\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.032784 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5wbk\" (UniqueName: \"kubernetes.io/projected/2e44c177-b87d-4ff6-80ca-672477fe9e94-kube-api-access-q5wbk\") pod \"nova-operator-controller-manager-569cc54c5-dzj5w\" (UID: \"2e44c177-b87d-4ff6-80ca-672477fe9e94\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.033081 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tk87\" (UniqueName: \"kubernetes.io/projected/309d4a2a-f738-4d2d-a28e-f361f762f997-kube-api-access-4tk87\") pod \"swift-operator-controller-manager-677c674df7-r2cxk\" (UID: \"309d4a2a-f738-4d2d-a28e-f361f762f997\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.033151 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q249w\" (UniqueName: \"kubernetes.io/projected/ccd465e8-6811-4865-9602-3dea8144cc01-kube-api-access-q249w\") pod \"mariadb-operator-controller-manager-658d4cdd5-c58lh\" (UID: \"ccd465e8-6811-4865-9602-3dea8144cc01\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.033184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7j4b\" (UniqueName: \"kubernetes.io/projected/464725b8-2734-43a4-a232-5db9bafed311-kube-api-access-w7j4b\") pod \"octavia-operator-controller-manager-5f4f55cb5c-cv97b\" (UID: \"464725b8-2734-43a4-a232-5db9bafed311\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.033243 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.033284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf95s\" (UniqueName: \"kubernetes.io/projected/5d4dea90-1696-4195-a0a0-71a3c9f3e328-kube-api-access-nf95s\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.047633 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.087702 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q249w\" (UniqueName: \"kubernetes.io/projected/ccd465e8-6811-4865-9602-3dea8144cc01-kube-api-access-q249w\") pod \"mariadb-operator-controller-manager-658d4cdd5-c58lh\" (UID: \"ccd465e8-6811-4865-9602-3dea8144cc01\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.091784 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rslhc\" (UniqueName: \"kubernetes.io/projected/6374b396-00ef-4aca-ac07-fd46982f23f1-kube-api-access-rslhc\") pod \"neutron-operator-controller-manager-776c5696bf-njrjb\" (UID: \"6374b396-00ef-4aca-ac07-fd46982f23f1\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.097316 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5wbk\" (UniqueName: \"kubernetes.io/projected/2e44c177-b87d-4ff6-80ca-672477fe9e94-kube-api-access-q5wbk\") pod \"nova-operator-controller-manager-569cc54c5-dzj5w\" (UID: \"2e44c177-b87d-4ff6-80ca-672477fe9e94\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.101590 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.103915 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.118569 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-hwv2s" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.122897 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.135045 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tk87\" (UniqueName: \"kubernetes.io/projected/309d4a2a-f738-4d2d-a28e-f361f762f997-kube-api-access-4tk87\") pod \"swift-operator-controller-manager-677c674df7-r2cxk\" (UID: \"309d4a2a-f738-4d2d-a28e-f361f762f997\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.136263 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7j4b\" (UniqueName: \"kubernetes.io/projected/464725b8-2734-43a4-a232-5db9bafed311-kube-api-access-w7j4b\") pod \"octavia-operator-controller-manager-5f4f55cb5c-cv97b\" (UID: \"464725b8-2734-43a4-a232-5db9bafed311\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.136347 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.136385 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf95s\" (UniqueName: \"kubernetes.io/projected/5d4dea90-1696-4195-a0a0-71a3c9f3e328-kube-api-access-nf95s\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.136500 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f48c2\" (UniqueName: \"kubernetes.io/projected/14346445-95be-488f-858c-44bf5b45c656-kube-api-access-f48c2\") pod \"ovn-operator-controller-manager-bbc5b68f9-xr747\" (UID: \"14346445-95be-488f-858c-44bf5b45c656\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.136565 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg8wk\" (UniqueName: \"kubernetes.io/projected/e3baa344-9dd8-48ed-8b6a-60ff9fbc181a-kube-api-access-hg8wk\") pod \"placement-operator-controller-manager-574d45c66c-tpnlt\" (UID: \"e3baa344-9dd8-48ed-8b6a-60ff9fbc181a\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.137715 4926 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.137809 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert podName:5d4dea90-1696-4195-a0a0-71a3c9f3e328 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:03.637756757 +0000 UTC m=+1044.006383090 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" (UID: "5d4dea90-1696-4195-a0a0-71a3c9f3e328") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.154673 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.159424 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.173641 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7j4b\" (UniqueName: \"kubernetes.io/projected/464725b8-2734-43a4-a232-5db9bafed311-kube-api-access-w7j4b\") pod \"octavia-operator-controller-manager-5f4f55cb5c-cv97b\" (UID: \"464725b8-2734-43a4-a232-5db9bafed311\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.176892 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f48c2\" (UniqueName: \"kubernetes.io/projected/14346445-95be-488f-858c-44bf5b45c656-kube-api-access-f48c2\") pod \"ovn-operator-controller-manager-bbc5b68f9-xr747\" (UID: \"14346445-95be-488f-858c-44bf5b45c656\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.177040 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tk87\" (UniqueName: \"kubernetes.io/projected/309d4a2a-f738-4d2d-a28e-f361f762f997-kube-api-access-4tk87\") pod \"swift-operator-controller-manager-677c674df7-r2cxk\" (UID: \"309d4a2a-f738-4d2d-a28e-f361f762f997\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.177490 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg8wk\" (UniqueName: \"kubernetes.io/projected/e3baa344-9dd8-48ed-8b6a-60ff9fbc181a-kube-api-access-hg8wk\") pod \"placement-operator-controller-manager-574d45c66c-tpnlt\" (UID: \"e3baa344-9dd8-48ed-8b6a-60ff9fbc181a\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.177628 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf95s\" (UniqueName: \"kubernetes.io/projected/5d4dea90-1696-4195-a0a0-71a3c9f3e328-kube-api-access-nf95s\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.189537 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.190376 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.193322 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-2tf46" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.193685 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.212644 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.231956 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.235765 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.236916 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.242373 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmp4h\" (UniqueName: \"kubernetes.io/projected/45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd-kube-api-access-fmp4h\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-z87mn\" (UID: \"45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.248775 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sm27l" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.252039 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.256355 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.275084 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.322376 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.330692 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.331202 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.331304 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.333588 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.334073 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-6t2jw" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.338535 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.343398 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.343463 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjm58\" (UniqueName: \"kubernetes.io/projected/682f3a0f-7437-455c-99e1-8b7cdb03328a-kube-api-access-pjm58\") pod \"test-operator-controller-manager-5c5cb9c4d7-25jrt\" (UID: \"682f3a0f-7437-455c-99e1-8b7cdb03328a\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.343487 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6s7t\" (UniqueName: \"kubernetes.io/projected/80f26a3f-ab1e-49b1-8843-6674d948a5cd-kube-api-access-d6s7t\") pod \"watcher-operator-controller-manager-6dd88c6f67-cn7fq\" (UID: \"80f26a3f-ab1e-49b1-8843-6674d948a5cd\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.343512 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmp4h\" (UniqueName: \"kubernetes.io/projected/45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd-kube-api-access-fmp4h\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-z87mn\" (UID: \"45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.343648 4926 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.343702 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert podName:fd35525c-7b73-49d1-a36c-c49d3bf933eb nodeName:}" failed. No retries permitted until 2026-03-12 18:20:04.343685984 +0000 UTC m=+1044.712312307 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert") pod "infra-operator-controller-manager-5995f4446f-kzl7p" (UID: "fd35525c-7b73-49d1-a36c-c49d3bf933eb") : secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.358728 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.360066 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.361035 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.365706 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-2q2p5" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.365929 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.372986 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmp4h\" (UniqueName: \"kubernetes.io/projected/45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd-kube-api-access-fmp4h\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-z87mn\" (UID: \"45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.406680 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.444240 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjm58\" (UniqueName: \"kubernetes.io/projected/682f3a0f-7437-455c-99e1-8b7cdb03328a-kube-api-access-pjm58\") pod \"test-operator-controller-manager-5c5cb9c4d7-25jrt\" (UID: \"682f3a0f-7437-455c-99e1-8b7cdb03328a\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.444280 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6s7t\" (UniqueName: \"kubernetes.io/projected/80f26a3f-ab1e-49b1-8843-6674d948a5cd-kube-api-access-d6s7t\") pod \"watcher-operator-controller-manager-6dd88c6f67-cn7fq\" (UID: \"80f26a3f-ab1e-49b1-8843-6674d948a5cd\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.444337 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdhjg\" (UniqueName: \"kubernetes.io/projected/fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b-kube-api-access-fdhjg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wqzk9\" (UID: \"fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.444410 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.445045 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q6pl\" (UniqueName: \"kubernetes.io/projected/495cec8d-a262-4d55-9ee5-6eebb10b6765-kube-api-access-6q6pl\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.445072 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.470972 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjm58\" (UniqueName: \"kubernetes.io/projected/682f3a0f-7437-455c-99e1-8b7cdb03328a-kube-api-access-pjm58\") pod \"test-operator-controller-manager-5c5cb9c4d7-25jrt\" (UID: \"682f3a0f-7437-455c-99e1-8b7cdb03328a\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.473983 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6s7t\" (UniqueName: \"kubernetes.io/projected/80f26a3f-ab1e-49b1-8843-6674d948a5cd-kube-api-access-d6s7t\") pod \"watcher-operator-controller-manager-6dd88c6f67-cn7fq\" (UID: \"80f26a3f-ab1e-49b1-8843-6674d948a5cd\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.546515 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdhjg\" (UniqueName: \"kubernetes.io/projected/fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b-kube-api-access-fdhjg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wqzk9\" (UID: \"fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.546802 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.546905 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q6pl\" (UniqueName: \"kubernetes.io/projected/495cec8d-a262-4d55-9ee5-6eebb10b6765-kube-api-access-6q6pl\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.546985 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.547169 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.547267 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:04.047252658 +0000 UTC m=+1044.415878991 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.547592 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.547698 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:04.04768845 +0000 UTC m=+1044.416314783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "metrics-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.569352 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q6pl\" (UniqueName: \"kubernetes.io/projected/495cec8d-a262-4d55-9ee5-6eebb10b6765-kube-api-access-6q6pl\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.570827 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdhjg\" (UniqueName: \"kubernetes.io/projected/fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b-kube-api-access-fdhjg\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wqzk9\" (UID: \"fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.577135 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.636918 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.648367 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.648592 4926 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: E0312 18:20:03.649650 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert podName:5d4dea90-1696-4195-a0a0-71a3c9f3e328 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:04.649619642 +0000 UTC m=+1045.018246055 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" (UID: "5d4dea90-1696-4195-a0a0-71a3c9f3e328") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.657171 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.693993 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.706902 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.738822 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf"] Mar 12 18:20:03 crc kubenswrapper[4926]: I0312 18:20:03.933348 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.055632 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.055994 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.056141 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.056209 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:05.056186822 +0000 UTC m=+1045.424813155 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "webhook-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.056522 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.056564 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:05.056551783 +0000 UTC m=+1045.425178116 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "metrics-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.308406 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx"] Mar 12 18:20:04 crc kubenswrapper[4926]: W0312 18:20:04.322782 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod065fe73a_651c_4cd3_b8d7_135617c51bbd.slice/crio-769059b055d102ba84effcc25a0d3adf60d1bd5e73e018efdcfab3e6a8f50679 WatchSource:0}: Error finding container 769059b055d102ba84effcc25a0d3adf60d1bd5e73e018efdcfab3e6a8f50679: Status 404 returned error can't find the container with id 769059b055d102ba84effcc25a0d3adf60d1bd5e73e018efdcfab3e6a8f50679 Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.328965 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.362118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.362396 4926 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.362473 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert podName:fd35525c-7b73-49d1-a36c-c49d3bf933eb nodeName:}" failed. No retries permitted until 2026-03-12 18:20:06.362432629 +0000 UTC m=+1046.731058962 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert") pod "infra-operator-controller-manager-5995f4446f-kzl7p" (UID: "fd35525c-7b73-49d1-a36c-c49d3bf933eb") : secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.362756 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-525z5"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.545698 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555660-cbx5b" Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.667979 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6s5b\" (UniqueName: \"kubernetes.io/projected/b50fb579-57d6-4029-a4f3-c8a3303bac4d-kube-api-access-f6s5b\") pod \"b50fb579-57d6-4029-a4f3-c8a3303bac4d\" (UID: \"b50fb579-57d6-4029-a4f3-c8a3303bac4d\") " Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.668304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.668594 4926 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: E0312 18:20:04.668693 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert podName:5d4dea90-1696-4195-a0a0-71a3c9f3e328 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:06.668668837 +0000 UTC m=+1047.037295210 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" (UID: "5d4dea90-1696-4195-a0a0-71a3c9f3e328") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.687319 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b50fb579-57d6-4029-a4f3-c8a3303bac4d-kube-api-access-f6s5b" (OuterVolumeSpecName: "kube-api-access-f6s5b") pod "b50fb579-57d6-4029-a4f3-c8a3303bac4d" (UID: "b50fb579-57d6-4029-a4f3-c8a3303bac4d"). InnerVolumeSpecName "kube-api-access-f6s5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.722970 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.729865 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.746634 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.770060 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6s5b\" (UniqueName: \"kubernetes.io/projected/b50fb579-57d6-4029-a4f3-c8a3303bac4d-kube-api-access-f6s5b\") on node \"crc\" DevicePath \"\"" Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.773063 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.792856 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.803039 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.875229 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" event={"ID":"b776be98-1352-43c6-8ee8-e31076b7d12b","Type":"ContainerStarted","Data":"3d8da108de485eedef583392ee33111fceb0e719a5535289dc724da90b0da38d"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.877456 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" event={"ID":"fea71415-42ac-4e77-ba9c-25170ccece27","Type":"ContainerStarted","Data":"0ac5f9503dca95dd3000648533056ba61503a5029e4d1cdeddcb1dde4d15f5be"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.878636 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" event={"ID":"4800992f-cfad-4a1c-94e5-79427f88c002","Type":"ContainerStarted","Data":"8b66ab767d4f939f6e4072b918286c560db1a63ee14df30b0cfacbef55a2a510"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.882290 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555660-cbx5b" event={"ID":"b50fb579-57d6-4029-a4f3-c8a3303bac4d","Type":"ContainerDied","Data":"f263435b763cc115efbb77597e4b93b8d2c4a1f2f6ba06fd3f2f757eb1164e16"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.882322 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f263435b763cc115efbb77597e4b93b8d2c4a1f2f6ba06fd3f2f757eb1164e16" Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.882367 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555660-cbx5b" Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.893600 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" event={"ID":"14346445-95be-488f-858c-44bf5b45c656","Type":"ContainerStarted","Data":"6cb8eda1b70336068b0d2ce8120252d67265b7566856a8302132942ab925b06d"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.897050 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" event={"ID":"065fe73a-651c-4cd3-b8d7-135617c51bbd","Type":"ContainerStarted","Data":"769059b055d102ba84effcc25a0d3adf60d1bd5e73e018efdcfab3e6a8f50679"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.898376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" event={"ID":"c62baaa0-d72b-4240-9dba-858bdf61d1b3","Type":"ContainerStarted","Data":"364c57258777706c1f0a24919ec22ef6537c98d949064c3415e910cc8c29cd6c"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.899405 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" event={"ID":"f112cb87-7454-41fa-a1e1-381d79f86247","Type":"ContainerStarted","Data":"51e730ea5972d666a55bf8d4c459c11da26f5c9bac76f90c3683958639006898"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.900781 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" event={"ID":"e38cf931-bbd6-4b47-bdf6-8a514d17d3d7","Type":"ContainerStarted","Data":"44f318f32ee118844455e111852282b7cf9d2b00c171ca1de24ab41a19fa9fe3"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.904854 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" event={"ID":"a41da562-a119-4785-95d0-eaf0970a99f4","Type":"ContainerStarted","Data":"ec458530e7483a1fa11b051ad6e9776ee400c828e8cc57d04fbddf74f8415400"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.907836 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" event={"ID":"e7b3fba0-ddaa-4cef-9df6-0683a92475cf","Type":"ContainerStarted","Data":"e5317c600ec77fbdf355fbc5e88cec64b1b8c074a90201367a8a16a527007d4b"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.909180 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" event={"ID":"464725b8-2734-43a4-a232-5db9bafed311","Type":"ContainerStarted","Data":"ed0dfb9c9bd787bf5584b3bf7cecc5cd59350cfffef8eddebbf70c4c9c22fd98"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.910606 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" event={"ID":"e3baa344-9dd8-48ed-8b6a-60ff9fbc181a","Type":"ContainerStarted","Data":"a61437a3a4f883b3927166fd4ef0d47de54e75d5f2abcd145622b9256b247c27"} Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.973078 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.982237 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.988764 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn"] Mar 12 18:20:04 crc kubenswrapper[4926]: I0312 18:20:04.995835 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb"] Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.001468 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq"] Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.008760 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt"] Mar 12 18:20:05 crc kubenswrapper[4926]: W0312 18:20:05.010138 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod682f3a0f_7437_455c_99e1_8b7cdb03328a.slice/crio-335cc127803561e2b305c4cec2665d2eebd4d9255df56259a72ffb8b0f0af07b WatchSource:0}: Error finding container 335cc127803561e2b305c4cec2665d2eebd4d9255df56259a72ffb8b0f0af07b: Status 404 returned error can't find the container with id 335cc127803561e2b305c4cec2665d2eebd4d9255df56259a72ffb8b0f0af07b Mar 12 18:20:05 crc kubenswrapper[4926]: W0312 18:20:05.010669 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc6165e6_3d8f_4ddd_b6ae_a1307f2c6b3b.slice/crio-5ef16cf6580abee89b6c12f326482775078e8159285ebfe8f969b5da392be107 WatchSource:0}: Error finding container 5ef16cf6580abee89b6c12f326482775078e8159285ebfe8f969b5da392be107: Status 404 returned error can't find the container with id 5ef16cf6580abee89b6c12f326482775078e8159285ebfe8f969b5da392be107 Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.010956 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w"] Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.015120 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh"] Mar 12 18:20:05 crc kubenswrapper[4926]: W0312 18:20:05.016131 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80f26a3f_ab1e_49b1_8843_6674d948a5cd.slice/crio-7fbaa0f112fbfa5bb70bb72a413114ef56396c426f8b3faacf7d85355d570657 WatchSource:0}: Error finding container 7fbaa0f112fbfa5bb70bb72a413114ef56396c426f8b3faacf7d85355d570657: Status 404 returned error can't find the container with id 7fbaa0f112fbfa5bb70bb72a413114ef56396c426f8b3faacf7d85355d570657 Mar 12 18:20:05 crc kubenswrapper[4926]: W0312 18:20:05.018779 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod309d4a2a_f738_4d2d_a28e_f361f762f997.slice/crio-2e657873af44236a401f737091b0b44d128edeb0486cf0aa500b55a6753caf8e WatchSource:0}: Error finding container 2e657873af44236a401f737091b0b44d128edeb0486cf0aa500b55a6753caf8e: Status 404 returned error can't find the container with id 2e657873af44236a401f737091b0b44d128edeb0486cf0aa500b55a6753caf8e Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.019387 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d6s7t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6dd88c6f67-cn7fq_openstack-operators(80f26a3f-ab1e-49b1-8843-6674d948a5cd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.020839 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" podUID="80f26a3f-ab1e-49b1-8843-6674d948a5cd" Mar 12 18:20:05 crc kubenswrapper[4926]: W0312 18:20:05.034534 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6374b396_00ef_4aca_ac07_fd46982f23f1.slice/crio-95746e7b99ab105d593ac84930ec3c380ba9152103a98c99c3a6a6f2bd8f80c3 WatchSource:0}: Error finding container 95746e7b99ab105d593ac84930ec3c380ba9152103a98c99c3a6a6f2bd8f80c3: Status 404 returned error can't find the container with id 95746e7b99ab105d593ac84930ec3c380ba9152103a98c99c3a6a6f2bd8f80c3 Mar 12 18:20:05 crc kubenswrapper[4926]: W0312 18:20:05.042941 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccd465e8_6811_4865_9602_3dea8144cc01.slice/crio-c9cb830c25538d7dd733478498e753da1234bed8d1846481094737779bdc41d0 WatchSource:0}: Error finding container c9cb830c25538d7dd733478498e753da1234bed8d1846481094737779bdc41d0: Status 404 returned error can't find the container with id c9cb830c25538d7dd733478498e753da1234bed8d1846481094737779bdc41d0 Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.043081 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4tk87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-677c674df7-r2cxk_openstack-operators(309d4a2a-f738-4d2d-a28e-f361f762f997): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.044222 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q5wbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-569cc54c5-dzj5w_openstack-operators(2e44c177-b87d-4ff6-80ca-672477fe9e94): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.044373 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rslhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-776c5696bf-njrjb_openstack-operators(6374b396-00ef-4aca-ac07-fd46982f23f1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.044460 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" podUID="309d4a2a-f738-4d2d-a28e-f361f762f997" Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.046043 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" podUID="6374b396-00ef-4aca-ac07-fd46982f23f1" Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.046096 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" podUID="2e44c177-b87d-4ff6-80ca-672477fe9e94" Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.074939 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.074996 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.075140 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.075198 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:07.075184765 +0000 UTC m=+1047.443811098 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "webhook-server-cert" not found Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.075224 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.075287 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:07.075270988 +0000 UTC m=+1047.443897321 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "metrics-server-cert" not found Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.618530 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555654-jjvcf"] Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.627150 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555654-jjvcf"] Mar 12 18:20:05 crc kubenswrapper[4926]: I0312 18:20:05.953624 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" event={"ID":"2e44c177-b87d-4ff6-80ca-672477fe9e94","Type":"ContainerStarted","Data":"3cc4bf1aa3b6fb6ba3ceb48192827b5a35da1ef79262d29fcc2f5c84013b48fa"} Mar 12 18:20:05 crc kubenswrapper[4926]: E0312 18:20:05.963609 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922\\\"\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" podUID="2e44c177-b87d-4ff6-80ca-672477fe9e94" Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.000216 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" event={"ID":"45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd","Type":"ContainerStarted","Data":"686314c6445feb8e08cad680a4dd637dfe9e9d1abfe8ec01f297b4f3f04b5879"} Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.043425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" event={"ID":"6374b396-00ef-4aca-ac07-fd46982f23f1","Type":"ContainerStarted","Data":"95746e7b99ab105d593ac84930ec3c380ba9152103a98c99c3a6a6f2bd8f80c3"} Mar 12 18:20:06 crc kubenswrapper[4926]: E0312 18:20:06.044890 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" podUID="6374b396-00ef-4aca-ac07-fd46982f23f1" Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.046039 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" event={"ID":"309d4a2a-f738-4d2d-a28e-f361f762f997","Type":"ContainerStarted","Data":"2e657873af44236a401f737091b0b44d128edeb0486cf0aa500b55a6753caf8e"} Mar 12 18:20:06 crc kubenswrapper[4926]: E0312 18:20:06.051647 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" podUID="309d4a2a-f738-4d2d-a28e-f361f762f997" Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.052391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" event={"ID":"80f26a3f-ab1e-49b1-8843-6674d948a5cd","Type":"ContainerStarted","Data":"7fbaa0f112fbfa5bb70bb72a413114ef56396c426f8b3faacf7d85355d570657"} Mar 12 18:20:06 crc kubenswrapper[4926]: E0312 18:20:06.053562 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" podUID="80f26a3f-ab1e-49b1-8843-6674d948a5cd" Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.057283 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" event={"ID":"ccd465e8-6811-4865-9602-3dea8144cc01","Type":"ContainerStarted","Data":"c9cb830c25538d7dd733478498e753da1234bed8d1846481094737779bdc41d0"} Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.058781 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" event={"ID":"682f3a0f-7437-455c-99e1-8b7cdb03328a","Type":"ContainerStarted","Data":"335cc127803561e2b305c4cec2665d2eebd4d9255df56259a72ffb8b0f0af07b"} Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.060007 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" event={"ID":"fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b","Type":"ContainerStarted","Data":"5ef16cf6580abee89b6c12f326482775078e8159285ebfe8f969b5da392be107"} Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.401093 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:06 crc kubenswrapper[4926]: E0312 18:20:06.401337 4926 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:06 crc kubenswrapper[4926]: E0312 18:20:06.401411 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert podName:fd35525c-7b73-49d1-a36c-c49d3bf933eb nodeName:}" failed. No retries permitted until 2026-03-12 18:20:10.401395886 +0000 UTC m=+1050.770022219 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert") pod "infra-operator-controller-manager-5995f4446f-kzl7p" (UID: "fd35525c-7b73-49d1-a36c-c49d3bf933eb") : secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.509037 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3e47fc3-fa77-4992-b1f8-dfff8b4d924e" path="/var/lib/kubelet/pods/b3e47fc3-fa77-4992-b1f8-dfff8b4d924e/volumes" Mar 12 18:20:06 crc kubenswrapper[4926]: I0312 18:20:06.714181 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:06 crc kubenswrapper[4926]: E0312 18:20:06.714473 4926 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:06 crc kubenswrapper[4926]: E0312 18:20:06.714768 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert podName:5d4dea90-1696-4195-a0a0-71a3c9f3e328 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:10.714693123 +0000 UTC m=+1051.083319456 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" (UID: "5d4dea90-1696-4195-a0a0-71a3c9f3e328") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.076711 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" podUID="80f26a3f-ab1e-49b1-8843-6674d948a5cd" Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.076725 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922\\\"\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" podUID="2e44c177-b87d-4ff6-80ca-672477fe9e94" Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.076788 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" podUID="6374b396-00ef-4aca-ac07-fd46982f23f1" Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.076851 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" podUID="309d4a2a-f738-4d2d-a28e-f361f762f997" Mar 12 18:20:07 crc kubenswrapper[4926]: I0312 18:20:07.120669 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:07 crc kubenswrapper[4926]: I0312 18:20:07.120804 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.121646 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.121688 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:11.121675235 +0000 UTC m=+1051.490301568 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "metrics-server-cert" not found Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.121724 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 18:20:07 crc kubenswrapper[4926]: E0312 18:20:07.121743 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:11.121737327 +0000 UTC m=+1051.490363660 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "webhook-server-cert" not found Mar 12 18:20:10 crc kubenswrapper[4926]: I0312 18:20:10.481460 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:10 crc kubenswrapper[4926]: E0312 18:20:10.481689 4926 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:10 crc kubenswrapper[4926]: E0312 18:20:10.481925 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert podName:fd35525c-7b73-49d1-a36c-c49d3bf933eb nodeName:}" failed. No retries permitted until 2026-03-12 18:20:18.481907359 +0000 UTC m=+1058.850533692 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert") pod "infra-operator-controller-manager-5995f4446f-kzl7p" (UID: "fd35525c-7b73-49d1-a36c-c49d3bf933eb") : secret "infra-operator-webhook-server-cert" not found Mar 12 18:20:10 crc kubenswrapper[4926]: I0312 18:20:10.786403 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:10 crc kubenswrapper[4926]: E0312 18:20:10.786592 4926 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:10 crc kubenswrapper[4926]: E0312 18:20:10.786668 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert podName:5d4dea90-1696-4195-a0a0-71a3c9f3e328 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:18.78665023 +0000 UTC m=+1059.155276563 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" (UID: "5d4dea90-1696-4195-a0a0-71a3c9f3e328") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 12 18:20:11 crc kubenswrapper[4926]: I0312 18:20:11.190838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:11 crc kubenswrapper[4926]: I0312 18:20:11.190894 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:11 crc kubenswrapper[4926]: E0312 18:20:11.191010 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 12 18:20:11 crc kubenswrapper[4926]: E0312 18:20:11.191056 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:19.191043171 +0000 UTC m=+1059.559669504 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "webhook-server-cert" not found Mar 12 18:20:11 crc kubenswrapper[4926]: E0312 18:20:11.191421 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 12 18:20:11 crc kubenswrapper[4926]: E0312 18:20:11.191474 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs podName:495cec8d-a262-4d55-9ee5-6eebb10b6765 nodeName:}" failed. No retries permitted until 2026-03-12 18:20:19.191463475 +0000 UTC m=+1059.560089818 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs") pod "openstack-operator-controller-manager-55976db558-2kgv6" (UID: "495cec8d-a262-4d55-9ee5-6eebb10b6765") : secret "metrics-server-cert" not found Mar 12 18:20:17 crc kubenswrapper[4926]: E0312 18:20:17.726573 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571" Mar 12 18:20:17 crc kubenswrapper[4926]: E0312 18:20:17.727389 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-w7j4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5f4f55cb5c-cv97b_openstack-operators(464725b8-2734-43a4-a232-5db9bafed311): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:20:17 crc kubenswrapper[4926]: E0312 18:20:17.728578 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" podUID="464725b8-2734-43a4-a232-5db9bafed311" Mar 12 18:20:18 crc kubenswrapper[4926]: E0312 18:20:18.166753 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" podUID="464725b8-2734-43a4-a232-5db9bafed311" Mar 12 18:20:18 crc kubenswrapper[4926]: E0312 18:20:18.304576 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" Mar 12 18:20:18 crc kubenswrapper[4926]: E0312 18:20:18.304795 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pjm58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-25jrt_openstack-operators(682f3a0f-7437-455c-99e1-8b7cdb03328a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:20:18 crc kubenswrapper[4926]: E0312 18:20:18.306104 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" podUID="682f3a0f-7437-455c-99e1-8b7cdb03328a" Mar 12 18:20:18 crc kubenswrapper[4926]: I0312 18:20:18.503341 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:18 crc kubenswrapper[4926]: I0312 18:20:18.510971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fd35525c-7b73-49d1-a36c-c49d3bf933eb-cert\") pod \"infra-operator-controller-manager-5995f4446f-kzl7p\" (UID: \"fd35525c-7b73-49d1-a36c-c49d3bf933eb\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:18 crc kubenswrapper[4926]: I0312 18:20:18.620356 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:18 crc kubenswrapper[4926]: I0312 18:20:18.811348 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:18 crc kubenswrapper[4926]: I0312 18:20:18.825141 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5d4dea90-1696-4195-a0a0-71a3c9f3e328-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk\" (UID: \"5d4dea90-1696-4195-a0a0-71a3c9f3e328\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:18 crc kubenswrapper[4926]: I0312 18:20:18.832262 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:18 crc kubenswrapper[4926]: E0312 18:20:18.989053 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f" Mar 12 18:20:18 crc kubenswrapper[4926]: E0312 18:20:18.989222 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d9sjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6bbb499bbc-7q8xr_openstack-operators(a41da562-a119-4785-95d0-eaf0970a99f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:20:18 crc kubenswrapper[4926]: E0312 18:20:18.991364 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" podUID="a41da562-a119-4785-95d0-eaf0970a99f4" Mar 12 18:20:19 crc kubenswrapper[4926]: E0312 18:20:19.171767 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" podUID="a41da562-a119-4785-95d0-eaf0970a99f4" Mar 12 18:20:19 crc kubenswrapper[4926]: E0312 18:20:19.171859 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" podUID="682f3a0f-7437-455c-99e1-8b7cdb03328a" Mar 12 18:20:19 crc kubenswrapper[4926]: I0312 18:20:19.217013 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:19 crc kubenswrapper[4926]: I0312 18:20:19.217170 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:19 crc kubenswrapper[4926]: I0312 18:20:19.220924 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-metrics-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:19 crc kubenswrapper[4926]: I0312 18:20:19.221002 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/495cec8d-a262-4d55-9ee5-6eebb10b6765-webhook-certs\") pod \"openstack-operator-controller-manager-55976db558-2kgv6\" (UID: \"495cec8d-a262-4d55-9ee5-6eebb10b6765\") " pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:19 crc kubenswrapper[4926]: I0312 18:20:19.272318 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:20 crc kubenswrapper[4926]: E0312 18:20:20.487117 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Mar 12 18:20:20 crc kubenswrapper[4926]: E0312 18:20:20.487623 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fdhjg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-wqzk9_openstack-operators(fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:20:20 crc kubenswrapper[4926]: E0312 18:20:20.489686 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" podUID="fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b" Mar 12 18:20:21 crc kubenswrapper[4926]: E0312 18:20:21.061134 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca" Mar 12 18:20:21 crc kubenswrapper[4926]: E0312 18:20:21.061377 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4pcf6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-684f77d66d-5f6c4_openstack-operators(c62baaa0-d72b-4240-9dba-858bdf61d1b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:20:21 crc kubenswrapper[4926]: E0312 18:20:21.063001 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" podUID="c62baaa0-d72b-4240-9dba-858bdf61d1b3" Mar 12 18:20:21 crc kubenswrapper[4926]: E0312 18:20:21.182573 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" podUID="c62baaa0-d72b-4240-9dba-858bdf61d1b3" Mar 12 18:20:21 crc kubenswrapper[4926]: E0312 18:20:21.182778 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" podUID="fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b" Mar 12 18:20:21 crc kubenswrapper[4926]: I0312 18:20:21.956777 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk"] Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.001174 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6"] Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.048105 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p"] Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.194531 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" event={"ID":"e38cf931-bbd6-4b47-bdf6-8a514d17d3d7","Type":"ContainerStarted","Data":"828c5c775ada2fe5f95234cafd67fc4668597e16ad44e77f67724f9598d381bf"} Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.194641 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.198499 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" event={"ID":"e7b3fba0-ddaa-4cef-9df6-0683a92475cf","Type":"ContainerStarted","Data":"defe8e630055cea1ed7e76af9b7480ffc04f09076c35545b6298469cfafe3415"} Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.198650 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.200408 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" event={"ID":"065fe73a-651c-4cd3-b8d7-135617c51bbd","Type":"ContainerStarted","Data":"1522e86e37c89182735cd000418f29a0cbe10a5e5e870d136dd2747a3dc542ef"} Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.200493 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.202503 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" event={"ID":"e3baa344-9dd8-48ed-8b6a-60ff9fbc181a","Type":"ContainerStarted","Data":"6cb7e85519441ebc424d3fc21a8e17cb17c9bb6c61b7f2a163ccc016db5dbf47"} Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.202627 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.213504 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" podStartSLOduration=4.004369284 podStartE2EDuration="20.213484733s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.827990784 +0000 UTC m=+1045.196617117" lastFinishedPulling="2026-03-12 18:20:21.037106233 +0000 UTC m=+1061.405732566" observedRunningTime="2026-03-12 18:20:22.213314727 +0000 UTC m=+1062.581941060" watchObservedRunningTime="2026-03-12 18:20:22.213484733 +0000 UTC m=+1062.582111066" Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.227792 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" podStartSLOduration=3.035423707 podStartE2EDuration="20.227779087s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:03.84433033 +0000 UTC m=+1044.212956663" lastFinishedPulling="2026-03-12 18:20:21.03668571 +0000 UTC m=+1061.405312043" observedRunningTime="2026-03-12 18:20:22.226942541 +0000 UTC m=+1062.595568874" watchObservedRunningTime="2026-03-12 18:20:22.227779087 +0000 UTC m=+1062.596405420" Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.246580 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" podStartSLOduration=3.537968062 podStartE2EDuration="20.246561591s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.329423282 +0000 UTC m=+1044.698049615" lastFinishedPulling="2026-03-12 18:20:21.038016821 +0000 UTC m=+1061.406643144" observedRunningTime="2026-03-12 18:20:22.242988691 +0000 UTC m=+1062.611615024" watchObservedRunningTime="2026-03-12 18:20:22.246561591 +0000 UTC m=+1062.615187924" Mar 12 18:20:22 crc kubenswrapper[4926]: I0312 18:20:22.257952 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" podStartSLOduration=3.563995602 podStartE2EDuration="20.257930105s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.344169091 +0000 UTC m=+1044.712795434" lastFinishedPulling="2026-03-12 18:20:21.038103604 +0000 UTC m=+1061.406729937" observedRunningTime="2026-03-12 18:20:22.256810071 +0000 UTC m=+1062.625436404" watchObservedRunningTime="2026-03-12 18:20:22.257930105 +0000 UTC m=+1062.626556448" Mar 12 18:20:22 crc kubenswrapper[4926]: W0312 18:20:22.759210 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d4dea90_1696_4195_a0a0_71a3c9f3e328.slice/crio-88b9fb6167625e93eee57ccc328dfab0817f506629002b33f245a06fcba6ce90 WatchSource:0}: Error finding container 88b9fb6167625e93eee57ccc328dfab0817f506629002b33f245a06fcba6ce90: Status 404 returned error can't find the container with id 88b9fb6167625e93eee57ccc328dfab0817f506629002b33f245a06fcba6ce90 Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.241751 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" event={"ID":"309d4a2a-f738-4d2d-a28e-f361f762f997","Type":"ContainerStarted","Data":"763feb0ffeb642cdca0bbd08795356fb1d89c52e6e0d42195168772ebf5643ff"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.242271 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.246662 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" event={"ID":"14346445-95be-488f-858c-44bf5b45c656","Type":"ContainerStarted","Data":"1e1e50fa2f078b7411a21bc28789c426d7e98f9c928fcaed0397b34bc5590e1a"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.247282 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.249737 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" event={"ID":"2e44c177-b87d-4ff6-80ca-672477fe9e94","Type":"ContainerStarted","Data":"06660c125bb692846ae880a8122d118f266ae4c6008193fb85fbd988d5404f66"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.250144 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.262864 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" podStartSLOduration=3.465089555 podStartE2EDuration="21.262822539s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.042964092 +0000 UTC m=+1045.411590425" lastFinishedPulling="2026-03-12 18:20:22.840697076 +0000 UTC m=+1063.209323409" observedRunningTime="2026-03-12 18:20:23.262688525 +0000 UTC m=+1063.631314858" watchObservedRunningTime="2026-03-12 18:20:23.262822539 +0000 UTC m=+1063.631448872" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.263849 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" event={"ID":"45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd","Type":"ContainerStarted","Data":"c0443003c4743aa7a8eadd5f18ce00657edab23715d985580dfd8099373775a9"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.264493 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.267425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" event={"ID":"f112cb87-7454-41fa-a1e1-381d79f86247","Type":"ContainerStarted","Data":"72543ac4a663a52b3e29b7d8ba5c55081c5c3f12c96fb35dc79e48ecb592054b"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.267559 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.284687 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" event={"ID":"fea71415-42ac-4e77-ba9c-25170ccece27","Type":"ContainerStarted","Data":"a48c53a7579a238f9cb007586cb237c48f90369196db63fd60dbc80fe597c890"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.285299 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.306188 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" event={"ID":"4800992f-cfad-4a1c-94e5-79427f88c002","Type":"ContainerStarted","Data":"c2e7fdb0fc833c08fda5b2daeee9daeb435751ae04920ae6c3e37e2b25bf03ee"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.306229 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.310671 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" event={"ID":"ccd465e8-6811-4865-9602-3dea8144cc01","Type":"ContainerStarted","Data":"6153011a4207e693ecf652341f8cbc71a87f22cba9fde5b2f3412942f4009725"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.311350 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.314815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" event={"ID":"fd35525c-7b73-49d1-a36c-c49d3bf933eb","Type":"ContainerStarted","Data":"7aeb0adefead3376b85330b09dcae3d11e44ab37d90aa71590a5de68302d9924"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.319190 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" podStartSLOduration=4.615689543 podStartE2EDuration="21.319176723s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.044068117 +0000 UTC m=+1045.412694450" lastFinishedPulling="2026-03-12 18:20:21.747555297 +0000 UTC m=+1062.116181630" observedRunningTime="2026-03-12 18:20:23.314822578 +0000 UTC m=+1063.683448921" watchObservedRunningTime="2026-03-12 18:20:23.319176723 +0000 UTC m=+1063.687803056" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.319964 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" podStartSLOduration=5.086988966 podStartE2EDuration="21.319957037s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.81242374 +0000 UTC m=+1045.181050073" lastFinishedPulling="2026-03-12 18:20:21.045391811 +0000 UTC m=+1061.414018144" observedRunningTime="2026-03-12 18:20:23.293290417 +0000 UTC m=+1063.661916750" watchObservedRunningTime="2026-03-12 18:20:23.319957037 +0000 UTC m=+1063.688583370" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.322745 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" event={"ID":"495cec8d-a262-4d55-9ee5-6eebb10b6765","Type":"ContainerStarted","Data":"685059c69032e95006cf858050966322f44a4eb6be7907d55e6e91f8dc07ee66"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.322782 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" event={"ID":"495cec8d-a262-4d55-9ee5-6eebb10b6765","Type":"ContainerStarted","Data":"8aa1ab094661eae0b82ab1807123f94c2047ccfbc663b32ef8677c123b3e8cfe"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.323499 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.329152 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" event={"ID":"b776be98-1352-43c6-8ee8-e31076b7d12b","Type":"ContainerStarted","Data":"db53c8de65025967409dd19d5adc0986b9c49d06c657ab39c391e6ced0dff218"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.329791 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.335848 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" podStartSLOduration=4.248416426 podStartE2EDuration="21.335835391s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:03.950133532 +0000 UTC m=+1044.318759865" lastFinishedPulling="2026-03-12 18:20:21.037552497 +0000 UTC m=+1061.406178830" observedRunningTime="2026-03-12 18:20:23.335243292 +0000 UTC m=+1063.703869635" watchObservedRunningTime="2026-03-12 18:20:23.335835391 +0000 UTC m=+1063.704461724" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.342053 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" event={"ID":"5d4dea90-1696-4195-a0a0-71a3c9f3e328","Type":"ContainerStarted","Data":"88b9fb6167625e93eee57ccc328dfab0817f506629002b33f245a06fcba6ce90"} Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.357772 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" podStartSLOduration=4.701462042 podStartE2EDuration="21.357755333s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.38140738 +0000 UTC m=+1044.750033713" lastFinishedPulling="2026-03-12 18:20:21.037700671 +0000 UTC m=+1061.406327004" observedRunningTime="2026-03-12 18:20:23.354696708 +0000 UTC m=+1063.723323041" watchObservedRunningTime="2026-03-12 18:20:23.357755333 +0000 UTC m=+1063.726381676" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.415764 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" podStartSLOduration=5.3691268149999996 podStartE2EDuration="21.415746978s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.005016322 +0000 UTC m=+1045.373642665" lastFinishedPulling="2026-03-12 18:20:21.051636495 +0000 UTC m=+1061.420262828" observedRunningTime="2026-03-12 18:20:23.387379225 +0000 UTC m=+1063.756005558" watchObservedRunningTime="2026-03-12 18:20:23.415746978 +0000 UTC m=+1063.784373311" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.444292 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" podStartSLOduration=5.444544481 podStartE2EDuration="21.444274955s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.059238169 +0000 UTC m=+1045.427864492" lastFinishedPulling="2026-03-12 18:20:21.058968633 +0000 UTC m=+1061.427594966" observedRunningTime="2026-03-12 18:20:23.417080109 +0000 UTC m=+1063.785706442" watchObservedRunningTime="2026-03-12 18:20:23.444274955 +0000 UTC m=+1063.812901288" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.448349 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" podStartSLOduration=4.261933277 podStartE2EDuration="21.448337921s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:03.85043351 +0000 UTC m=+1044.219059843" lastFinishedPulling="2026-03-12 18:20:21.036838154 +0000 UTC m=+1061.405464487" observedRunningTime="2026-03-12 18:20:23.447728132 +0000 UTC m=+1063.816354465" watchObservedRunningTime="2026-03-12 18:20:23.448337921 +0000 UTC m=+1063.816964254" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.564087 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" podStartSLOduration=5.311507463 podStartE2EDuration="21.564070333s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.812800412 +0000 UTC m=+1045.181426745" lastFinishedPulling="2026-03-12 18:20:21.065363282 +0000 UTC m=+1061.433989615" observedRunningTime="2026-03-12 18:20:23.507939066 +0000 UTC m=+1063.876565399" watchObservedRunningTime="2026-03-12 18:20:23.564070333 +0000 UTC m=+1063.932696666" Mar 12 18:20:23 crc kubenswrapper[4926]: I0312 18:20:23.565546 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" podStartSLOduration=20.565542408 podStartE2EDuration="20.565542408s" podCreationTimestamp="2026-03-12 18:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:20:23.56111466 +0000 UTC m=+1063.929740993" watchObservedRunningTime="2026-03-12 18:20:23.565542408 +0000 UTC m=+1063.934168731" Mar 12 18:20:26 crc kubenswrapper[4926]: I0312 18:20:26.817991 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:20:26 crc kubenswrapper[4926]: I0312 18:20:26.818572 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.412899 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" event={"ID":"fd35525c-7b73-49d1-a36c-c49d3bf933eb","Type":"ContainerStarted","Data":"6b651cdbce774b8a347381e80f13dd9f73cc82aaa5c2ba84ea09ce1f53cf5d0b"} Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.413185 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.414855 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" event={"ID":"6374b396-00ef-4aca-ac07-fd46982f23f1","Type":"ContainerStarted","Data":"944c57984d00334e897efd2e8fe4538356b0be1c8b8ec63402fd0cf921161440"} Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.415112 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.416217 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" event={"ID":"5d4dea90-1696-4195-a0a0-71a3c9f3e328","Type":"ContainerStarted","Data":"ac118a798d9939b08069f9d91a941181a79f28ac4bcd090bdaefb227e03d138c"} Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.416281 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.417709 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" event={"ID":"80f26a3f-ab1e-49b1-8843-6674d948a5cd","Type":"ContainerStarted","Data":"61ae93568ede1eb4e0b871691c777333b050ef442c1812ff765143573d2c5149"} Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.417874 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.436651 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" podStartSLOduration=21.072744135 podStartE2EDuration="26.436629897s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:22.80933297 +0000 UTC m=+1063.177959313" lastFinishedPulling="2026-03-12 18:20:28.173218732 +0000 UTC m=+1068.541845075" observedRunningTime="2026-03-12 18:20:28.429280169 +0000 UTC m=+1068.797906542" watchObservedRunningTime="2026-03-12 18:20:28.436629897 +0000 UTC m=+1068.805256230" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.460128 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" podStartSLOduration=21.088855876 podStartE2EDuration="26.460110108s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:22.761599095 +0000 UTC m=+1063.130225428" lastFinishedPulling="2026-03-12 18:20:28.132853317 +0000 UTC m=+1068.501479660" observedRunningTime="2026-03-12 18:20:28.451925484 +0000 UTC m=+1068.820551817" watchObservedRunningTime="2026-03-12 18:20:28.460110108 +0000 UTC m=+1068.828736441" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.475293 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" podStartSLOduration=3.376682504 podStartE2EDuration="26.475275299s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.044307414 +0000 UTC m=+1045.412933747" lastFinishedPulling="2026-03-12 18:20:28.142900199 +0000 UTC m=+1068.511526542" observedRunningTime="2026-03-12 18:20:28.471287126 +0000 UTC m=+1068.839913459" watchObservedRunningTime="2026-03-12 18:20:28.475275299 +0000 UTC m=+1068.843901642" Mar 12 18:20:28 crc kubenswrapper[4926]: I0312 18:20:28.488307 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" podStartSLOduration=2.371981368 podStartE2EDuration="25.488290435s" podCreationTimestamp="2026-03-12 18:20:03 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.019274805 +0000 UTC m=+1045.387901138" lastFinishedPulling="2026-03-12 18:20:28.135583842 +0000 UTC m=+1068.504210205" observedRunningTime="2026-03-12 18:20:28.486817669 +0000 UTC m=+1068.855444002" watchObservedRunningTime="2026-03-12 18:20:28.488290435 +0000 UTC m=+1068.856916768" Mar 12 18:20:29 crc kubenswrapper[4926]: I0312 18:20:29.285776 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-55976db558-2kgv6" Mar 12 18:20:32 crc kubenswrapper[4926]: I0312 18:20:32.768508 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-ssmx7" Mar 12 18:20:32 crc kubenswrapper[4926]: I0312 18:20:32.788977 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-zlsrd" Mar 12 18:20:32 crc kubenswrapper[4926]: I0312 18:20:32.820411 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-clngf" Mar 12 18:20:32 crc kubenswrapper[4926]: I0312 18:20:32.883275 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-m6sxh" Mar 12 18:20:32 crc kubenswrapper[4926]: I0312 18:20:32.886060 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-525z5" Mar 12 18:20:32 crc kubenswrapper[4926]: I0312 18:20:32.915001 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-p76mx" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.161815 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-f96w7" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.216356 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-tpnlt" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.236636 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-c58lh" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.262118 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-677c674df7-r2cxk" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.353777 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-njrjb" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.368664 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-dzj5w" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.415767 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-xr747" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.579562 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-z87mn" Mar 12 18:20:33 crc kubenswrapper[4926]: I0312 18:20:33.659754 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-cn7fq" Mar 12 18:20:38 crc kubenswrapper[4926]: I0312 18:20:38.628910 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-kzl7p" Mar 12 18:20:38 crc kubenswrapper[4926]: I0312 18:20:38.840526 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.544148 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" event={"ID":"682f3a0f-7437-455c-99e1-8b7cdb03328a","Type":"ContainerStarted","Data":"e6e51b327fb4785b9196fc01b7c7230b6d836342a4107acead6dbb71e7f13361"} Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.544896 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" event={"ID":"fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b","Type":"ContainerStarted","Data":"50c65420b38ce4755cb1a02016ba35111df1c92ad7f069eeba953c8cb113d774"} Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.545191 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.546416 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" event={"ID":"c62baaa0-d72b-4240-9dba-858bdf61d1b3","Type":"ContainerStarted","Data":"7a1b9c63ce02ac15ed7d321e38b223df41144d7b29531cda327c0d6a8edc5ea7"} Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.546687 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.547470 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" event={"ID":"464725b8-2734-43a4-a232-5db9bafed311","Type":"ContainerStarted","Data":"9dad9f033bfa07f0de5d8fdcf103a401f35d903a42ba20049335b13790cf0a34"} Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.547796 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.549049 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" event={"ID":"a41da562-a119-4785-95d0-eaf0970a99f4","Type":"ContainerStarted","Data":"8214da748fc5915f21ee97850d91b201196f611f9ddc3bb6fb8cb58347eeabf4"} Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.549515 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.574837 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" podStartSLOduration=4.020561858 podStartE2EDuration="40.574818626s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.770659461 +0000 UTC m=+1045.139285794" lastFinishedPulling="2026-03-12 18:20:41.324916229 +0000 UTC m=+1081.693542562" observedRunningTime="2026-03-12 18:20:42.573286969 +0000 UTC m=+1082.941913312" watchObservedRunningTime="2026-03-12 18:20:42.574818626 +0000 UTC m=+1082.943444969" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.592787 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" podStartSLOduration=4.28515098 podStartE2EDuration="40.592767485s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.017716606 +0000 UTC m=+1045.386342939" lastFinishedPulling="2026-03-12 18:20:41.325333071 +0000 UTC m=+1081.693959444" observedRunningTime="2026-03-12 18:20:42.560770449 +0000 UTC m=+1082.929396782" watchObservedRunningTime="2026-03-12 18:20:42.592767485 +0000 UTC m=+1082.961393818" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.606899 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wqzk9" podStartSLOduration=3.299419974 podStartE2EDuration="39.606881304s" podCreationTimestamp="2026-03-12 18:20:03 +0000 UTC" firstStartedPulling="2026-03-12 18:20:05.018221902 +0000 UTC m=+1045.386848235" lastFinishedPulling="2026-03-12 18:20:41.325683212 +0000 UTC m=+1081.694309565" observedRunningTime="2026-03-12 18:20:42.601817205 +0000 UTC m=+1082.970443538" watchObservedRunningTime="2026-03-12 18:20:42.606881304 +0000 UTC m=+1082.975507637" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.622615 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" podStartSLOduration=4.128823964 podStartE2EDuration="40.622598732s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.829978685 +0000 UTC m=+1045.198605018" lastFinishedPulling="2026-03-12 18:20:41.323753423 +0000 UTC m=+1081.692379786" observedRunningTime="2026-03-12 18:20:42.618836055 +0000 UTC m=+1082.987462388" watchObservedRunningTime="2026-03-12 18:20:42.622598732 +0000 UTC m=+1082.991225065" Mar 12 18:20:42 crc kubenswrapper[4926]: I0312 18:20:42.640534 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" podStartSLOduration=4.148124116 podStartE2EDuration="40.64051137s" podCreationTimestamp="2026-03-12 18:20:02 +0000 UTC" firstStartedPulling="2026-03-12 18:20:04.831493263 +0000 UTC m=+1045.200119596" lastFinishedPulling="2026-03-12 18:20:41.323880517 +0000 UTC m=+1081.692506850" observedRunningTime="2026-03-12 18:20:42.635580326 +0000 UTC m=+1083.004206659" watchObservedRunningTime="2026-03-12 18:20:42.64051137 +0000 UTC m=+1083.009137723" Mar 12 18:20:53 crc kubenswrapper[4926]: I0312 18:20:53.052102 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7q8xr" Mar 12 18:20:53 crc kubenswrapper[4926]: I0312 18:20:53.159294 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-5f6c4" Mar 12 18:20:53 crc kubenswrapper[4926]: I0312 18:20:53.362325 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-cv97b" Mar 12 18:20:53 crc kubenswrapper[4926]: I0312 18:20:53.647465 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-25jrt" Mar 12 18:20:56 crc kubenswrapper[4926]: I0312 18:20:56.817958 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:20:56 crc kubenswrapper[4926]: I0312 18:20:56.818332 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:21:01 crc kubenswrapper[4926]: I0312 18:21:01.361183 4926 scope.go:117] "RemoveContainer" containerID="5d29d1a82fd23f75a9296b7891f008517b3e7945af33e9449e2dfcb52711dd8e" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.977162 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9hlv4"] Mar 12 18:21:12 crc kubenswrapper[4926]: E0312 18:21:12.977865 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50fb579-57d6-4029-a4f3-c8a3303bac4d" containerName="oc" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.977877 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50fb579-57d6-4029-a4f3-c8a3303bac4d" containerName="oc" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.978023 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b50fb579-57d6-4029-a4f3-c8a3303bac4d" containerName="oc" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.978698 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.982700 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.982709 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.982700 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.983060 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ggplz" Mar 12 18:21:12 crc kubenswrapper[4926]: I0312 18:21:12.992662 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9hlv4"] Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.020673 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2x544"] Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.022728 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.032055 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.040992 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2x544"] Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.088062 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-config\") pod \"dnsmasq-dns-675f4bcbfc-9hlv4\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.088342 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvwd8\" (UniqueName: \"kubernetes.io/projected/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-kube-api-access-mvwd8\") pod \"dnsmasq-dns-675f4bcbfc-9hlv4\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.189225 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x94ql\" (UniqueName: \"kubernetes.io/projected/757e1395-7d78-496a-bd4e-eb7a13b90d10-kube-api-access-x94ql\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.189285 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-config\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.189334 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-config\") pod \"dnsmasq-dns-675f4bcbfc-9hlv4\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.189363 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.189395 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvwd8\" (UniqueName: \"kubernetes.io/projected/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-kube-api-access-mvwd8\") pod \"dnsmasq-dns-675f4bcbfc-9hlv4\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.190322 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-config\") pod \"dnsmasq-dns-675f4bcbfc-9hlv4\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.208698 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvwd8\" (UniqueName: \"kubernetes.io/projected/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-kube-api-access-mvwd8\") pod \"dnsmasq-dns-675f4bcbfc-9hlv4\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.290866 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-config\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.290948 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.290992 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x94ql\" (UniqueName: \"kubernetes.io/projected/757e1395-7d78-496a-bd4e-eb7a13b90d10-kube-api-access-x94ql\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.291919 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.292784 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-config\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.296564 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.314192 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x94ql\" (UniqueName: \"kubernetes.io/projected/757e1395-7d78-496a-bd4e-eb7a13b90d10-kube-api-access-x94ql\") pod \"dnsmasq-dns-78dd6ddcc-2x544\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.353748 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.554693 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9hlv4"] Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.634034 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2x544"] Mar 12 18:21:13 crc kubenswrapper[4926]: W0312 18:21:13.635620 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod757e1395_7d78_496a_bd4e_eb7a13b90d10.slice/crio-8ced1e2cf1a0fabd93eb5ebb045e1e613152c80826e96ac3f9497d2c59523b9b WatchSource:0}: Error finding container 8ced1e2cf1a0fabd93eb5ebb045e1e613152c80826e96ac3f9497d2c59523b9b: Status 404 returned error can't find the container with id 8ced1e2cf1a0fabd93eb5ebb045e1e613152c80826e96ac3f9497d2c59523b9b Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.848076 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" event={"ID":"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f","Type":"ContainerStarted","Data":"1cba9386f004d3a4567f81b3b21065a63a088e2f0cf17a6228bba62f9f3a3e10"} Mar 12 18:21:13 crc kubenswrapper[4926]: I0312 18:21:13.849361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" event={"ID":"757e1395-7d78-496a-bd4e-eb7a13b90d10","Type":"ContainerStarted","Data":"8ced1e2cf1a0fabd93eb5ebb045e1e613152c80826e96ac3f9497d2c59523b9b"} Mar 12 18:21:15 crc kubenswrapper[4926]: I0312 18:21:15.961046 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9hlv4"] Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.003026 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-px8d4"] Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.004122 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.037980 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-px8d4"] Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.134492 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwd9r\" (UniqueName: \"kubernetes.io/projected/6e12f792-0fbe-4996-a6ec-272e5108dd33-kube-api-access-jwd9r\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.134588 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.134664 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-config\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.236130 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.236521 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-config\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.236626 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwd9r\" (UniqueName: \"kubernetes.io/projected/6e12f792-0fbe-4996-a6ec-272e5108dd33-kube-api-access-jwd9r\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.237332 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.237676 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-config\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.259002 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwd9r\" (UniqueName: \"kubernetes.io/projected/6e12f792-0fbe-4996-a6ec-272e5108dd33-kube-api-access-jwd9r\") pod \"dnsmasq-dns-5ccc8479f9-px8d4\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.275933 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2x544"] Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.304865 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lk7rj"] Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.306318 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.320669 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lk7rj"] Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.323019 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.440284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6jt2\" (UniqueName: \"kubernetes.io/projected/5241fdef-63ed-416f-9bf0-c004597cc099-kube-api-access-r6jt2\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.440359 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-config\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.440395 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.541555 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6jt2\" (UniqueName: \"kubernetes.io/projected/5241fdef-63ed-416f-9bf0-c004597cc099-kube-api-access-r6jt2\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.541911 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-config\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.541960 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.542846 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.543128 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-config\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.560802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6jt2\" (UniqueName: \"kubernetes.io/projected/5241fdef-63ed-416f-9bf0-c004597cc099-kube-api-access-r6jt2\") pod \"dnsmasq-dns-57d769cc4f-lk7rj\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.615533 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-px8d4"] Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.634739 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:16 crc kubenswrapper[4926]: W0312 18:21:16.640593 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e12f792_0fbe_4996_a6ec_272e5108dd33.slice/crio-9f0b5cf18f405df691a53e33e3f957de2dcab02a1304d75c9067e908e6be9be3 WatchSource:0}: Error finding container 9f0b5cf18f405df691a53e33e3f957de2dcab02a1304d75c9067e908e6be9be3: Status 404 returned error can't find the container with id 9f0b5cf18f405df691a53e33e3f957de2dcab02a1304d75c9067e908e6be9be3 Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.876027 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" event={"ID":"6e12f792-0fbe-4996-a6ec-272e5108dd33","Type":"ContainerStarted","Data":"9f0b5cf18f405df691a53e33e3f957de2dcab02a1304d75c9067e908e6be9be3"} Mar 12 18:21:16 crc kubenswrapper[4926]: I0312 18:21:16.916091 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lk7rj"] Mar 12 18:21:16 crc kubenswrapper[4926]: W0312 18:21:16.927780 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5241fdef_63ed_416f_9bf0_c004597cc099.slice/crio-22b5f71bb0db465dea314a1f0530175669b6d30c754c5e44cf8c2d8ddfdcad3d WatchSource:0}: Error finding container 22b5f71bb0db465dea314a1f0530175669b6d30c754c5e44cf8c2d8ddfdcad3d: Status 404 returned error can't find the container with id 22b5f71bb0db465dea314a1f0530175669b6d30c754c5e44cf8c2d8ddfdcad3d Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.146214 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.147814 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.159978 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.160232 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.160405 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.160581 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.160722 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.160872 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.161074 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-r4knd" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.166556 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250210 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250254 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kzhl\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-kube-api-access-8kzhl\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250283 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250304 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250332 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250487 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250562 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250613 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250686 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250759 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.250793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.351829 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.352100 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.352127 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.352219 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353316 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353390 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kzhl\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-kube-api-access-8kzhl\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353424 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353459 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353485 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353513 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353539 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353562 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.353991 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.354349 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.355882 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.356458 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.356997 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.358307 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.362316 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.362865 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.365302 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.383162 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kzhl\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-kube-api-access-8kzhl\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.393842 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.434737 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.438680 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.445283 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.445920 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.446597 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.446642 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.446797 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-k928p" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.446960 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.447266 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.447941 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.503431 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556316 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-config-data\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556370 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7kq2\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-kube-api-access-x7kq2\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556417 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556452 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556472 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c04aaec-485d-492f-8c24-e6860d9c78f7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556668 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556714 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556781 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556851 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556883 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.556932 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c04aaec-485d-492f-8c24-e6860d9c78f7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658741 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-config-data\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658789 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7kq2\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-kube-api-access-x7kq2\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658824 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658847 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c04aaec-485d-492f-8c24-e6860d9c78f7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658902 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658922 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658954 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.658982 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.659003 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.659025 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c04aaec-485d-492f-8c24-e6860d9c78f7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.659337 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.659595 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.660367 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.659334 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.666252 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.668952 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.669456 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-config-data\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.670673 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c04aaec-485d-492f-8c24-e6860d9c78f7-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.672105 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.676041 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7kq2\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-kube-api-access-x7kq2\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.684992 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.685104 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c04aaec-485d-492f-8c24-e6860d9c78f7-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.769165 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 18:21:17 crc kubenswrapper[4926]: I0312 18:21:17.883421 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" event={"ID":"5241fdef-63ed-416f-9bf0-c004597cc099","Type":"ContainerStarted","Data":"22b5f71bb0db465dea314a1f0530175669b6d30c754c5e44cf8c2d8ddfdcad3d"} Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.327475 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.329233 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.335700 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9fnd5" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.341629 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.342320 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.342737 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.348850 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.354299 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.377773 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.377818 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.377845 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-kolla-config\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.377891 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-config-data-default\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.377908 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.377929 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v6wb\" (UniqueName: \"kubernetes.io/projected/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-kube-api-access-5v6wb\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.377990 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.378020 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.479316 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.479489 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.480700 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.486576 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.486738 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.486782 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.486823 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-kolla-config\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.486861 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-config-data-default\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.486881 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.486903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v6wb\" (UniqueName: \"kubernetes.io/projected/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-kube-api-access-5v6wb\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.488841 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-kolla-config\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.492836 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.493193 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.494075 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-config-data-default\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.501784 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.512012 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v6wb\" (UniqueName: \"kubernetes.io/projected/1ee00086-3c8a-4f3a-a5d5-9590715a8b95-kube-api-access-5v6wb\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.520975 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"1ee00086-3c8a-4f3a-a5d5-9590715a8b95\") " pod="openstack/openstack-galera-0" Mar 12 18:21:18 crc kubenswrapper[4926]: I0312 18:21:18.653087 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.725561 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.726934 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.730462 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.730730 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.731486 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.731655 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-fdjl6" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.741396 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804135 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804377 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/94d38c16-a6c9-44ed-a49e-398dc34b92ce-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804423 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d38c16-a6c9-44ed-a49e-398dc34b92ce-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804484 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804543 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/94d38c16-a6c9-44ed-a49e-398dc34b92ce-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.804596 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llxz7\" (UniqueName: \"kubernetes.io/projected/94d38c16-a6c9-44ed-a49e-398dc34b92ce-kube-api-access-llxz7\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.905980 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.906036 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.906106 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.906144 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/94d38c16-a6c9-44ed-a49e-398dc34b92ce-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.906169 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d38c16-a6c9-44ed-a49e-398dc34b92ce-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.906245 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.906715 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/94d38c16-a6c9-44ed-a49e-398dc34b92ce-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.906961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.907039 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/94d38c16-a6c9-44ed-a49e-398dc34b92ce-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.907096 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llxz7\" (UniqueName: \"kubernetes.io/projected/94d38c16-a6c9-44ed-a49e-398dc34b92ce-kube-api-access-llxz7\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.907818 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.908097 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.908519 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/94d38c16-a6c9-44ed-a49e-398dc34b92ce-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.910913 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/94d38c16-a6c9-44ed-a49e-398dc34b92ce-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.912153 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94d38c16-a6c9-44ed-a49e-398dc34b92ce-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.927482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llxz7\" (UniqueName: \"kubernetes.io/projected/94d38c16-a6c9-44ed-a49e-398dc34b92ce-kube-api-access-llxz7\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:19 crc kubenswrapper[4926]: I0312 18:21:19.933551 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"94d38c16-a6c9-44ed-a49e-398dc34b92ce\") " pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.048850 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.235282 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.236190 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.237837 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-g9zp2" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.244920 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.247228 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.247497 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.319109 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/69083379-a7d7-4876-9955-497420eab579-memcached-tls-certs\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.319274 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/69083379-a7d7-4876-9955-497420eab579-kolla-config\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.319326 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69083379-a7d7-4876-9955-497420eab579-config-data\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.319354 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27297\" (UniqueName: \"kubernetes.io/projected/69083379-a7d7-4876-9955-497420eab579-kube-api-access-27297\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.319506 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69083379-a7d7-4876-9955-497420eab579-combined-ca-bundle\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.421954 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27297\" (UniqueName: \"kubernetes.io/projected/69083379-a7d7-4876-9955-497420eab579-kube-api-access-27297\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.422017 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69083379-a7d7-4876-9955-497420eab579-combined-ca-bundle\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.422073 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/69083379-a7d7-4876-9955-497420eab579-memcached-tls-certs\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.422125 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/69083379-a7d7-4876-9955-497420eab579-kolla-config\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.422148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69083379-a7d7-4876-9955-497420eab579-config-data\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.422861 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69083379-a7d7-4876-9955-497420eab579-config-data\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.424190 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/69083379-a7d7-4876-9955-497420eab579-kolla-config\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.426418 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/69083379-a7d7-4876-9955-497420eab579-memcached-tls-certs\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.439420 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27297\" (UniqueName: \"kubernetes.io/projected/69083379-a7d7-4876-9955-497420eab579-kube-api-access-27297\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.443159 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69083379-a7d7-4876-9955-497420eab579-combined-ca-bundle\") pod \"memcached-0\" (UID: \"69083379-a7d7-4876-9955-497420eab579\") " pod="openstack/memcached-0" Mar 12 18:21:20 crc kubenswrapper[4926]: I0312 18:21:20.551269 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.368911 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.370090 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.373736 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5578n" Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.383480 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.454950 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bjbq\" (UniqueName: \"kubernetes.io/projected/3343d19e-07d3-4de8-954a-f7e31aa8279f-kube-api-access-6bjbq\") pod \"kube-state-metrics-0\" (UID: \"3343d19e-07d3-4de8-954a-f7e31aa8279f\") " pod="openstack/kube-state-metrics-0" Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.556659 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bjbq\" (UniqueName: \"kubernetes.io/projected/3343d19e-07d3-4de8-954a-f7e31aa8279f-kube-api-access-6bjbq\") pod \"kube-state-metrics-0\" (UID: \"3343d19e-07d3-4de8-954a-f7e31aa8279f\") " pod="openstack/kube-state-metrics-0" Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.577202 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bjbq\" (UniqueName: \"kubernetes.io/projected/3343d19e-07d3-4de8-954a-f7e31aa8279f-kube-api-access-6bjbq\") pod \"kube-state-metrics-0\" (UID: \"3343d19e-07d3-4de8-954a-f7e31aa8279f\") " pod="openstack/kube-state-metrics-0" Mar 12 18:21:22 crc kubenswrapper[4926]: I0312 18:21:22.701181 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.838582 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sfwpr"] Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.839983 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.853504 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfwpr"] Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.855490 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.855858 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-64822" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.856025 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.880574 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6znbv"] Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.882576 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.900954 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6znbv"] Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.926616 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-run\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.926675 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-etc-ovs\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.926710 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-run-ovn\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.926741 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-lib\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.926962 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-log\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927023 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f19fbb7-ea3b-437a-a634-498e6a593ef6-ovn-controller-tls-certs\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927054 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-log-ovn\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927098 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30dfa384-92a5-49cf-9793-60478855264f-scripts\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927121 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-run\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927137 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f19fbb7-ea3b-437a-a634-498e6a593ef6-scripts\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927165 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56ctw\" (UniqueName: \"kubernetes.io/projected/8f19fbb7-ea3b-437a-a634-498e6a593ef6-kube-api-access-56ctw\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927250 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpdpt\" (UniqueName: \"kubernetes.io/projected/30dfa384-92a5-49cf-9793-60478855264f-kube-api-access-tpdpt\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.927273 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f19fbb7-ea3b-437a-a634-498e6a593ef6-combined-ca-bundle\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.946648 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.948133 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.954747 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.955142 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-s27qn" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.955284 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.955418 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.955497 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 12 18:21:25 crc kubenswrapper[4926]: I0312 18:21:25.960632 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.028162 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f19fbb7-ea3b-437a-a634-498e6a593ef6-ovn-controller-tls-certs\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029255 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-log-ovn\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029286 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c3288572-a9dc-4f96-8535-b05f6f22855b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029306 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30dfa384-92a5-49cf-9793-60478855264f-scripts\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029360 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f19fbb7-ea3b-437a-a634-498e6a593ef6-scripts\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-run\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029493 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56ctw\" (UniqueName: \"kubernetes.io/projected/8f19fbb7-ea3b-437a-a634-498e6a593ef6-kube-api-access-56ctw\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029533 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029587 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpdpt\" (UniqueName: \"kubernetes.io/projected/30dfa384-92a5-49cf-9793-60478855264f-kube-api-access-tpdpt\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029608 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f19fbb7-ea3b-437a-a634-498e6a593ef6-combined-ca-bundle\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029652 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029698 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3288572-a9dc-4f96-8535-b05f6f22855b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029763 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-run\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029816 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-etc-ovs\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029853 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-run-ovn\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029904 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-lib\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029977 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.029987 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-log-ovn\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030010 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chrgc\" (UniqueName: \"kubernetes.io/projected/c3288572-a9dc-4f96-8535-b05f6f22855b-kube-api-access-chrgc\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030068 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3288572-a9dc-4f96-8535-b05f6f22855b-config\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030100 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-log\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030063 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-run\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030314 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-log\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030396 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-run-ovn\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030450 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8f19fbb7-ea3b-437a-a634-498e6a593ef6-var-run\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030573 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-etc-ovs\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.030651 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/30dfa384-92a5-49cf-9793-60478855264f-var-lib\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.031312 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30dfa384-92a5-49cf-9793-60478855264f-scripts\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.031482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f19fbb7-ea3b-437a-a634-498e6a593ef6-scripts\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.033764 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f19fbb7-ea3b-437a-a634-498e6a593ef6-ovn-controller-tls-certs\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.034424 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f19fbb7-ea3b-437a-a634-498e6a593ef6-combined-ca-bundle\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.048094 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56ctw\" (UniqueName: \"kubernetes.io/projected/8f19fbb7-ea3b-437a-a634-498e6a593ef6-kube-api-access-56ctw\") pod \"ovn-controller-sfwpr\" (UID: \"8f19fbb7-ea3b-437a-a634-498e6a593ef6\") " pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.048415 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpdpt\" (UniqueName: \"kubernetes.io/projected/30dfa384-92a5-49cf-9793-60478855264f-kube-api-access-tpdpt\") pod \"ovn-controller-ovs-6znbv\" (UID: \"30dfa384-92a5-49cf-9793-60478855264f\") " pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131411 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chrgc\" (UniqueName: \"kubernetes.io/projected/c3288572-a9dc-4f96-8535-b05f6f22855b-kube-api-access-chrgc\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131430 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3288572-a9dc-4f96-8535-b05f6f22855b-config\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131509 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c3288572-a9dc-4f96-8535-b05f6f22855b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131528 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131571 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131596 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131617 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3288572-a9dc-4f96-8535-b05f6f22855b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.131796 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.132477 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c3288572-a9dc-4f96-8535-b05f6f22855b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.132671 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3288572-a9dc-4f96-8535-b05f6f22855b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.132684 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3288572-a9dc-4f96-8535-b05f6f22855b-config\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.135728 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.145744 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.146059 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3288572-a9dc-4f96-8535-b05f6f22855b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.148045 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chrgc\" (UniqueName: \"kubernetes.io/projected/c3288572-a9dc-4f96-8535-b05f6f22855b-kube-api-access-chrgc\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.161116 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c3288572-a9dc-4f96-8535-b05f6f22855b\") " pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.178566 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.207971 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.273637 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.818087 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.818168 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.818237 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.819151 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"10c4816f4e2fc4ce2bc2183a633d9bc53980639515bfce0cf198e862b133fadb"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.819306 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://10c4816f4e2fc4ce2bc2183a633d9bc53980639515bfce0cf198e862b133fadb" gracePeriod=600 Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.965768 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="10c4816f4e2fc4ce2bc2183a633d9bc53980639515bfce0cf198e862b133fadb" exitCode=0 Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.965845 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"10c4816f4e2fc4ce2bc2183a633d9bc53980639515bfce0cf198e862b133fadb"} Mar 12 18:21:26 crc kubenswrapper[4926]: I0312 18:21:26.966034 4926 scope.go:117] "RemoveContainer" containerID="a397bef079b1410b3294983dad25ada9109b1a0eac364c78c0ff4aeeccdf38ed" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.649566 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.650911 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.653386 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-hl59j" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.657182 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.657304 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.657360 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.674589 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685072 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8926ccac-b553-4e37-bbcb-96e3b00c1cab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685167 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685196 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685259 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdv8t\" (UniqueName: \"kubernetes.io/projected/8926ccac-b553-4e37-bbcb-96e3b00c1cab-kube-api-access-sdv8t\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685406 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8926ccac-b553-4e37-bbcb-96e3b00c1cab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685565 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.685642 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8926ccac-b553-4e37-bbcb-96e3b00c1cab-config\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787537 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdv8t\" (UniqueName: \"kubernetes.io/projected/8926ccac-b553-4e37-bbcb-96e3b00c1cab-kube-api-access-sdv8t\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787638 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787677 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8926ccac-b553-4e37-bbcb-96e3b00c1cab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787739 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787774 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8926ccac-b553-4e37-bbcb-96e3b00c1cab-config\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787813 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8926ccac-b553-4e37-bbcb-96e3b00c1cab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787907 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.787973 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.788300 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.790408 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/8926ccac-b553-4e37-bbcb-96e3b00c1cab-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.792941 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8926ccac-b553-4e37-bbcb-96e3b00c1cab-config\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.793014 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8926ccac-b553-4e37-bbcb-96e3b00c1cab-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.801021 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.801244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.806576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8926ccac-b553-4e37-bbcb-96e3b00c1cab-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.807701 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdv8t\" (UniqueName: \"kubernetes.io/projected/8926ccac-b553-4e37-bbcb-96e3b00c1cab-kube-api-access-sdv8t\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.809283 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"8926ccac-b553-4e37-bbcb-96e3b00c1cab\") " pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:29 crc kubenswrapper[4926]: I0312 18:21:29.972430 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:30 crc kubenswrapper[4926]: E0312 18:21:30.239177 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 12 18:21:30 crc kubenswrapper[4926]: E0312 18:21:30.239346 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x94ql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-2x544_openstack(757e1395-7d78-496a-bd4e-eb7a13b90d10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:21:30 crc kubenswrapper[4926]: E0312 18:21:30.240581 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" podUID="757e1395-7d78-496a-bd4e-eb7a13b90d10" Mar 12 18:21:30 crc kubenswrapper[4926]: E0312 18:21:30.354573 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 12 18:21:30 crc kubenswrapper[4926]: E0312 18:21:30.354988 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mvwd8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-9hlv4_openstack(b74b0a6f-8e6a-4be5-8c67-7473b7c0041f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:21:30 crc kubenswrapper[4926]: E0312 18:21:30.356290 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" podUID="b74b0a6f-8e6a-4be5-8c67-7473b7c0041f" Mar 12 18:21:30 crc kubenswrapper[4926]: I0312 18:21:30.998007 4926 generic.go:334] "Generic (PLEG): container finished" podID="5241fdef-63ed-416f-9bf0-c004597cc099" containerID="14fbc530f0eccf6bba2b3c2c222ca781991938a69c4f20b33818d33df7d57e2d" exitCode=0 Mar 12 18:21:30 crc kubenswrapper[4926]: I0312 18:21:30.998503 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" event={"ID":"5241fdef-63ed-416f-9bf0-c004597cc099","Type":"ContainerDied","Data":"14fbc530f0eccf6bba2b3c2c222ca781991938a69c4f20b33818d33df7d57e2d"} Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.013131 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"9728bd6132bdd9ab31a71d0a44779a02f515c39e712bb7cc4f8a85610efe739f"} Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.019566 4926 generic.go:334] "Generic (PLEG): container finished" podID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerID="70b08923dd3c9c83aaab8247e1a7a0c78b75d9c1349d70ce24d38f2bd6316ae1" exitCode=0 Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.019691 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" event={"ID":"6e12f792-0fbe-4996-a6ec-272e5108dd33","Type":"ContainerDied","Data":"70b08923dd3c9c83aaab8247e1a7a0c78b75d9c1349d70ce24d38f2bd6316ae1"} Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.079375 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.131617 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.142105 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.177910 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.210903 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.226305 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfwpr"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.259935 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.332192 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.532384 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.537360 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.685975 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 12 18:21:31 crc kubenswrapper[4926]: W0312 18:21:31.694540 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8926ccac_b553_4e37_bbcb_96e3b00c1cab.slice/crio-0e9f3a68e6e59617bb832dfb4b84662a80a7305b65003215be030597cbad60c8 WatchSource:0}: Error finding container 0e9f3a68e6e59617bb832dfb4b84662a80a7305b65003215be030597cbad60c8: Status 404 returned error can't find the container with id 0e9f3a68e6e59617bb832dfb4b84662a80a7305b65003215be030597cbad60c8 Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.733285 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-dns-svc\") pod \"757e1395-7d78-496a-bd4e-eb7a13b90d10\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.733781 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x94ql\" (UniqueName: \"kubernetes.io/projected/757e1395-7d78-496a-bd4e-eb7a13b90d10-kube-api-access-x94ql\") pod \"757e1395-7d78-496a-bd4e-eb7a13b90d10\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.733890 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "757e1395-7d78-496a-bd4e-eb7a13b90d10" (UID: "757e1395-7d78-496a-bd4e-eb7a13b90d10"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.733921 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvwd8\" (UniqueName: \"kubernetes.io/projected/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-kube-api-access-mvwd8\") pod \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.733966 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-config\") pod \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\" (UID: \"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f\") " Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.734004 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-config\") pod \"757e1395-7d78-496a-bd4e-eb7a13b90d10\" (UID: \"757e1395-7d78-496a-bd4e-eb7a13b90d10\") " Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.734376 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.734433 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-config" (OuterVolumeSpecName: "config") pod "757e1395-7d78-496a-bd4e-eb7a13b90d10" (UID: "757e1395-7d78-496a-bd4e-eb7a13b90d10"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.734485 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-config" (OuterVolumeSpecName: "config") pod "b74b0a6f-8e6a-4be5-8c67-7473b7c0041f" (UID: "b74b0a6f-8e6a-4be5-8c67-7473b7c0041f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.739957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-kube-api-access-mvwd8" (OuterVolumeSpecName: "kube-api-access-mvwd8") pod "b74b0a6f-8e6a-4be5-8c67-7473b7c0041f" (UID: "b74b0a6f-8e6a-4be5-8c67-7473b7c0041f"). InnerVolumeSpecName "kube-api-access-mvwd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.740585 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/757e1395-7d78-496a-bd4e-eb7a13b90d10-kube-api-access-x94ql" (OuterVolumeSpecName: "kube-api-access-x94ql") pod "757e1395-7d78-496a-bd4e-eb7a13b90d10" (UID: "757e1395-7d78-496a-bd4e-eb7a13b90d10"). InnerVolumeSpecName "kube-api-access-x94ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.790538 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6znbv"] Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.836124 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x94ql\" (UniqueName: \"kubernetes.io/projected/757e1395-7d78-496a-bd4e-eb7a13b90d10-kube-api-access-x94ql\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.836155 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvwd8\" (UniqueName: \"kubernetes.io/projected/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-kube-api-access-mvwd8\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.836166 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:31 crc kubenswrapper[4926]: I0312 18:21:31.836176 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/757e1395-7d78-496a-bd4e-eb7a13b90d10-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:31 crc kubenswrapper[4926]: W0312 18:21:31.875687 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30dfa384_92a5_49cf_9793_60478855264f.slice/crio-01c671ef539d352b60786d84428aff26871c5405e87f77c3ee1ad41b2dca46ff WatchSource:0}: Error finding container 01c671ef539d352b60786d84428aff26871c5405e87f77c3ee1ad41b2dca46ff: Status 404 returned error can't find the container with id 01c671ef539d352b60786d84428aff26871c5405e87f77c3ee1ad41b2dca46ff Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.031801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c3288572-a9dc-4f96-8535-b05f6f22855b","Type":"ContainerStarted","Data":"a78b120c7cf2870625a84495d8e3a18d8ff80a83d9bd025f6f535e76a6f9b7bf"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.043365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" event={"ID":"b74b0a6f-8e6a-4be5-8c67-7473b7c0041f","Type":"ContainerDied","Data":"1cba9386f004d3a4567f81b3b21065a63a088e2f0cf17a6228bba62f9f3a3e10"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.043575 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-9hlv4" Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.058730 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" event={"ID":"6e12f792-0fbe-4996-a6ec-272e5108dd33","Type":"ContainerStarted","Data":"2f4a53f9672845382051e0baaa943aad6fd054f0217c765f521ce6c2186e70ad"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.058907 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.066031 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" event={"ID":"5241fdef-63ed-416f-9bf0-c004597cc099","Type":"ContainerStarted","Data":"40595f3e3c0223d3ed3f1599101a865666d9ee1a40df2b9373fe5655ddc620fa"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.066120 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.067552 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8926ccac-b553-4e37-bbcb-96e3b00c1cab","Type":"ContainerStarted","Data":"0e9f3a68e6e59617bb832dfb4b84662a80a7305b65003215be030597cbad60c8"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.069262 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6znbv" event={"ID":"30dfa384-92a5-49cf-9793-60478855264f","Type":"ContainerStarted","Data":"01c671ef539d352b60786d84428aff26871c5405e87f77c3ee1ad41b2dca46ff"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.070821 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.070864 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2x544" event={"ID":"757e1395-7d78-496a-bd4e-eb7a13b90d10","Type":"ContainerDied","Data":"8ced1e2cf1a0fabd93eb5ebb045e1e613152c80826e96ac3f9497d2c59523b9b"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.074331 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"94d38c16-a6c9-44ed-a49e-398dc34b92ce","Type":"ContainerStarted","Data":"8cffc6f2cf22e065c646327d7e2376f07621924b60d95b70b11c502492ded6be"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.075514 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06f09c04-6c8d-4c47-a0a5-59def6ebbf94","Type":"ContainerStarted","Data":"851881376da52e23e91e40413019b432ba8d1a58a75cb35642884721899d9a50"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.091697 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9c04aaec-485d-492f-8c24-e6860d9c78f7","Type":"ContainerStarted","Data":"1dfe0defccd325a5f2aba559c2e683f83b8337f73cdd57e5ea2b609807eb38bf"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.102021 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" podStartSLOduration=3.294161679 podStartE2EDuration="17.101998319s" podCreationTimestamp="2026-03-12 18:21:15 +0000 UTC" firstStartedPulling="2026-03-12 18:21:16.647341772 +0000 UTC m=+1117.015968105" lastFinishedPulling="2026-03-12 18:21:30.455178392 +0000 UTC m=+1130.823804745" observedRunningTime="2026-03-12 18:21:32.084191124 +0000 UTC m=+1132.452817457" watchObservedRunningTime="2026-03-12 18:21:32.101998319 +0000 UTC m=+1132.470624652" Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.109721 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"69083379-a7d7-4876-9955-497420eab579","Type":"ContainerStarted","Data":"4038b17eb45e5fa7b20e5c814863f1f8b28f82cb2cfecff14d8f152dce6de6b8"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.111235 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3343d19e-07d3-4de8-954a-f7e31aa8279f","Type":"ContainerStarted","Data":"a5a3fb3c5e3c756132a6d36939e2dc50de833c191b34382d6e423a84d04ecf45"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.117600 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" podStartSLOduration=2.600653134 podStartE2EDuration="16.117578673s" podCreationTimestamp="2026-03-12 18:21:16 +0000 UTC" firstStartedPulling="2026-03-12 18:21:16.936092206 +0000 UTC m=+1117.304718529" lastFinishedPulling="2026-03-12 18:21:30.453017735 +0000 UTC m=+1130.821644068" observedRunningTime="2026-03-12 18:21:32.108322856 +0000 UTC m=+1132.476949179" watchObservedRunningTime="2026-03-12 18:21:32.117578673 +0000 UTC m=+1132.486205006" Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.119561 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr" event={"ID":"8f19fbb7-ea3b-437a-a634-498e6a593ef6","Type":"ContainerStarted","Data":"0dee60d04f094592a5abba55f5c41b4de3a7242ba688e0eefb63ed7c3283abf5"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.126018 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1ee00086-3c8a-4f3a-a5d5-9590715a8b95","Type":"ContainerStarted","Data":"9927d7adb7f8c599b3e1dbf5bd1405228ef1744ac07a01bf6d24cd343a94144e"} Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.169502 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9hlv4"] Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.175337 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-9hlv4"] Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.194461 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2x544"] Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.206969 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2x544"] Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.500109 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="757e1395-7d78-496a-bd4e-eb7a13b90d10" path="/var/lib/kubelet/pods/757e1395-7d78-496a-bd4e-eb7a13b90d10/volumes" Mar 12 18:21:32 crc kubenswrapper[4926]: I0312 18:21:32.500566 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b74b0a6f-8e6a-4be5-8c67-7473b7c0041f" path="/var/lib/kubelet/pods/b74b0a6f-8e6a-4be5-8c67-7473b7c0041f/volumes" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.759922 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-mszqk"] Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.793046 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mszqk"] Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.793160 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.795554 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.888483 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.888522 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-ovs-rundir\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.888545 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-combined-ca-bundle\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.888570 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-ovn-rundir\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.888611 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-config\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.888633 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhxd2\" (UniqueName: \"kubernetes.io/projected/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-kube-api-access-nhxd2\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.928737 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-px8d4"] Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.966335 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6csg8"] Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.967552 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.969940 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.984334 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6csg8"] Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990281 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-ovs-rundir\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990330 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990352 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-combined-ca-bundle\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-ovn-rundir\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990422 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990460 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-config\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990484 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhxd2\" (UniqueName: \"kubernetes.io/projected/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-kube-api-access-nhxd2\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990529 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-config\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990562 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28bxt\" (UniqueName: \"kubernetes.io/projected/10777956-1f18-4bd0-9790-aae020334e2c-kube-api-access-28bxt\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990598 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.990968 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-ovs-rundir\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.991508 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-ovn-rundir\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.992272 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-config\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:33 crc kubenswrapper[4926]: I0312 18:21:33.999133 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-combined-ca-bundle\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.008979 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.018141 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhxd2\" (UniqueName: \"kubernetes.io/projected/b4a45633-8ac7-497f-a8d7-2b7a3dad35bc-kube-api-access-nhxd2\") pod \"ovn-controller-metrics-mszqk\" (UID: \"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc\") " pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.081835 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lk7rj"] Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.092225 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-config\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.092304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28bxt\" (UniqueName: \"kubernetes.io/projected/10777956-1f18-4bd0-9790-aae020334e2c-kube-api-access-28bxt\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.092364 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.092453 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.093417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.093459 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-config\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.093458 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.124333 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-htjb7"] Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.127103 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.127951 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28bxt\" (UniqueName: \"kubernetes.io/projected/10777956-1f18-4bd0-9790-aae020334e2c-kube-api-access-28bxt\") pod \"dnsmasq-dns-7fd796d7df-6csg8\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.130555 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mszqk" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.133884 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.144949 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" containerName="dnsmasq-dns" containerID="cri-o://40595f3e3c0223d3ed3f1599101a865666d9ee1a40df2b9373fe5655ddc620fa" gracePeriod=10 Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.145289 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerName="dnsmasq-dns" containerID="cri-o://2f4a53f9672845382051e0baaa943aad6fd054f0217c765f521ce6c2186e70ad" gracePeriod=10 Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.148713 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-htjb7"] Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.195142 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.195179 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.195213 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.195264 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7c2g\" (UniqueName: \"kubernetes.io/projected/7919b4ef-74fc-472e-9a5e-04216cc51ae5-kube-api-access-l7c2g\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.195462 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-config\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.287072 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.297067 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-config\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.297148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.297172 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.297204 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.297230 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7c2g\" (UniqueName: \"kubernetes.io/projected/7919b4ef-74fc-472e-9a5e-04216cc51ae5-kube-api-access-l7c2g\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.298201 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.298243 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.298259 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.299466 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-config\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.317216 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7c2g\" (UniqueName: \"kubernetes.io/projected/7919b4ef-74fc-472e-9a5e-04216cc51ae5-kube-api-access-l7c2g\") pod \"dnsmasq-dns-86db49b7ff-htjb7\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:34 crc kubenswrapper[4926]: I0312 18:21:34.460706 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:35 crc kubenswrapper[4926]: I0312 18:21:35.164865 4926 generic.go:334] "Generic (PLEG): container finished" podID="5241fdef-63ed-416f-9bf0-c004597cc099" containerID="40595f3e3c0223d3ed3f1599101a865666d9ee1a40df2b9373fe5655ddc620fa" exitCode=0 Mar 12 18:21:35 crc kubenswrapper[4926]: I0312 18:21:35.164965 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" event={"ID":"5241fdef-63ed-416f-9bf0-c004597cc099","Type":"ContainerDied","Data":"40595f3e3c0223d3ed3f1599101a865666d9ee1a40df2b9373fe5655ddc620fa"} Mar 12 18:21:35 crc kubenswrapper[4926]: I0312 18:21:35.167498 4926 generic.go:334] "Generic (PLEG): container finished" podID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerID="2f4a53f9672845382051e0baaa943aad6fd054f0217c765f521ce6c2186e70ad" exitCode=0 Mar 12 18:21:35 crc kubenswrapper[4926]: I0312 18:21:35.167524 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" event={"ID":"6e12f792-0fbe-4996-a6ec-272e5108dd33","Type":"ContainerDied","Data":"2f4a53f9672845382051e0baaa943aad6fd054f0217c765f521ce6c2186e70ad"} Mar 12 18:21:36 crc kubenswrapper[4926]: I0312 18:21:36.326714 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.99:5353: connect: connection refused" Mar 12 18:21:40 crc kubenswrapper[4926]: E0312 18:21:40.390801 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = reading blob sha256:49c8ff4b1361fbc5f5e8704c41aa7b56c23335467834ce27b461e2430f9aa48d: Digest did not match, expected sha256:49c8ff4b1361fbc5f5e8704c41aa7b56c23335467834ce27b461e2430f9aa48d, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" image="quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified" Mar 12 18:21:40 crc kubenswrapper[4926]: E0312 18:21:40.391527 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-controller,Image:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,Command:[ovn-controller --pidfile unix:/run/openvswitch/db.sock --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n87h665h5cfh55dh5chdfhdch55fh7ch56dhc8h9bh5d8hb7hc9h54ch597h66fh7fh645h5ddh56ch554h5b8h5f6h5cfh5f6h5b8hfh5bdh58dh5ddq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-run,ReadOnly:false,MountPath:/var/run/openvswitch,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-run-ovn,ReadOnly:false,MountPath:/var/run/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:var-log-ovn,ReadOnly:false,MountPath:/var/log/ovn,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-controller-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-56ctw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_liveness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/ovn_controller_readiness.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:30,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/usr/share/ovn/scripts/ovn-ctl stop_controller],},HTTPGet:nil,TCPSocket:nil,Sleep:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN SYS_ADMIN SYS_NICE],Drop:[],},Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-controller-sfwpr_openstack(8f19fbb7-ea3b-437a-a634-498e6a593ef6): ErrImagePull: reading blob sha256:49c8ff4b1361fbc5f5e8704c41aa7b56c23335467834ce27b461e2430f9aa48d: Digest did not match, expected sha256:49c8ff4b1361fbc5f5e8704c41aa7b56c23335467834ce27b461e2430f9aa48d, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" logger="UnhandledError" Mar 12 18:21:40 crc kubenswrapper[4926]: E0312 18:21:40.393040 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ErrImagePull: \"reading blob sha256:49c8ff4b1361fbc5f5e8704c41aa7b56c23335467834ce27b461e2430f9aa48d: Digest did not match, expected sha256:49c8ff4b1361fbc5f5e8704c41aa7b56c23335467834ce27b461e2430f9aa48d, got sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855\"" pod="openstack/ovn-controller-sfwpr" podUID="8f19fbb7-ea3b-437a-a634-498e6a593ef6" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.408601 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.603841 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-dns-svc\") pod \"5241fdef-63ed-416f-9bf0-c004597cc099\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.604244 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6jt2\" (UniqueName: \"kubernetes.io/projected/5241fdef-63ed-416f-9bf0-c004597cc099-kube-api-access-r6jt2\") pod \"5241fdef-63ed-416f-9bf0-c004597cc099\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.608059 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-config\") pod \"5241fdef-63ed-416f-9bf0-c004597cc099\" (UID: \"5241fdef-63ed-416f-9bf0-c004597cc099\") " Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.609622 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5241fdef-63ed-416f-9bf0-c004597cc099-kube-api-access-r6jt2" (OuterVolumeSpecName: "kube-api-access-r6jt2") pod "5241fdef-63ed-416f-9bf0-c004597cc099" (UID: "5241fdef-63ed-416f-9bf0-c004597cc099"). InnerVolumeSpecName "kube-api-access-r6jt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.699951 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-config" (OuterVolumeSpecName: "config") pod "5241fdef-63ed-416f-9bf0-c004597cc099" (UID: "5241fdef-63ed-416f-9bf0-c004597cc099"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.700106 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5241fdef-63ed-416f-9bf0-c004597cc099" (UID: "5241fdef-63ed-416f-9bf0-c004597cc099"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.707712 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.710144 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.710183 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6jt2\" (UniqueName: \"kubernetes.io/projected/5241fdef-63ed-416f-9bf0-c004597cc099-kube-api-access-r6jt2\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.710193 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5241fdef-63ed-416f-9bf0-c004597cc099-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.811805 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-config\") pod \"6e12f792-0fbe-4996-a6ec-272e5108dd33\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.813324 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwd9r\" (UniqueName: \"kubernetes.io/projected/6e12f792-0fbe-4996-a6ec-272e5108dd33-kube-api-access-jwd9r\") pod \"6e12f792-0fbe-4996-a6ec-272e5108dd33\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.813565 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-dns-svc\") pod \"6e12f792-0fbe-4996-a6ec-272e5108dd33\" (UID: \"6e12f792-0fbe-4996-a6ec-272e5108dd33\") " Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.821947 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e12f792-0fbe-4996-a6ec-272e5108dd33-kube-api-access-jwd9r" (OuterVolumeSpecName: "kube-api-access-jwd9r") pod "6e12f792-0fbe-4996-a6ec-272e5108dd33" (UID: "6e12f792-0fbe-4996-a6ec-272e5108dd33"). InnerVolumeSpecName "kube-api-access-jwd9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.841489 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e12f792-0fbe-4996-a6ec-272e5108dd33" (UID: "6e12f792-0fbe-4996-a6ec-272e5108dd33"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.848113 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-config" (OuterVolumeSpecName: "config") pod "6e12f792-0fbe-4996-a6ec-272e5108dd33" (UID: "6e12f792-0fbe-4996-a6ec-272e5108dd33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.915734 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.915759 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e12f792-0fbe-4996-a6ec-272e5108dd33-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:40 crc kubenswrapper[4926]: I0312 18:21:40.915790 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwd9r\" (UniqueName: \"kubernetes.io/projected/6e12f792-0fbe-4996-a6ec-272e5108dd33-kube-api-access-jwd9r\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.229933 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" event={"ID":"5241fdef-63ed-416f-9bf0-c004597cc099","Type":"ContainerDied","Data":"22b5f71bb0db465dea314a1f0530175669b6d30c754c5e44cf8c2d8ddfdcad3d"} Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.229969 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.230038 4926 scope.go:117] "RemoveContainer" containerID="40595f3e3c0223d3ed3f1599101a865666d9ee1a40df2b9373fe5655ddc620fa" Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.234618 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" event={"ID":"6e12f792-0fbe-4996-a6ec-272e5108dd33","Type":"ContainerDied","Data":"9f0b5cf18f405df691a53e33e3f957de2dcab02a1304d75c9067e908e6be9be3"} Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.234754 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-px8d4" Mar 12 18:21:41 crc kubenswrapper[4926]: E0312 18:21:41.235808 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-controller\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified\\\"\"" pod="openstack/ovn-controller-sfwpr" podUID="8f19fbb7-ea3b-437a-a634-498e6a593ef6" Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.278681 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lk7rj"] Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.290539 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-lk7rj"] Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.299386 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-px8d4"] Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.307385 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-px8d4"] Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.637138 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-lk7rj" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.100:5353: i/o timeout" Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.745878 4926 scope.go:117] "RemoveContainer" containerID="14fbc530f0eccf6bba2b3c2c222ca781991938a69c4f20b33818d33df7d57e2d" Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.804380 4926 scope.go:117] "RemoveContainer" containerID="2f4a53f9672845382051e0baaa943aad6fd054f0217c765f521ce6c2186e70ad" Mar 12 18:21:41 crc kubenswrapper[4926]: I0312 18:21:41.861963 4926 scope.go:117] "RemoveContainer" containerID="70b08923dd3c9c83aaab8247e1a7a0c78b75d9c1349d70ce24d38f2bd6316ae1" Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.128349 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mszqk"] Mar 12 18:21:42 crc kubenswrapper[4926]: W0312 18:21:42.140714 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4a45633_8ac7_497f_a8d7_2b7a3dad35bc.slice/crio-f7c4dce115367598ad2159524b93cf71ec6949a14754d95aea2017f89097389d WatchSource:0}: Error finding container f7c4dce115367598ad2159524b93cf71ec6949a14754d95aea2017f89097389d: Status 404 returned error can't find the container with id f7c4dce115367598ad2159524b93cf71ec6949a14754d95aea2017f89097389d Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.243794 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6csg8"] Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.246997 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c3288572-a9dc-4f96-8535-b05f6f22855b","Type":"ContainerStarted","Data":"781ea0a760f08140b9883c5fed3a171df1919c1af90c679e6f40c64e3a5fa919"} Mar 12 18:21:42 crc kubenswrapper[4926]: W0312 18:21:42.250681 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10777956_1f18_4bd0_9790_aae020334e2c.slice/crio-1fc31c043cd931246845186f093dded25733b7a182735715b314acd79d3a77b3 WatchSource:0}: Error finding container 1fc31c043cd931246845186f093dded25733b7a182735715b314acd79d3a77b3: Status 404 returned error can't find the container with id 1fc31c043cd931246845186f093dded25733b7a182735715b314acd79d3a77b3 Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.250815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1ee00086-3c8a-4f3a-a5d5-9590715a8b95","Type":"ContainerStarted","Data":"cfea673c23f1afedb464a97e3d967b55a20c7c92aa61908df2cb8c1bc70bc1ba"} Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.253065 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8926ccac-b553-4e37-bbcb-96e3b00c1cab","Type":"ContainerStarted","Data":"8715b23c355f385951aad7ac8c2327f6b2a3947620ff78fcfcac3b66d8b5d8ed"} Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.257843 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mszqk" event={"ID":"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc","Type":"ContainerStarted","Data":"f7c4dce115367598ad2159524b93cf71ec6949a14754d95aea2017f89097389d"} Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.259739 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"94d38c16-a6c9-44ed-a49e-398dc34b92ce","Type":"ContainerStarted","Data":"fc66a1cd90ab8ffc8ce1b57a7c53af70932339daf452369540b2057b973e9dcc"} Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.261260 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"69083379-a7d7-4876-9955-497420eab579","Type":"ContainerStarted","Data":"78e04a3036a523b939ce014c96a8e4153242d2617ecfecaceb5bcfc558e634db"} Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.261662 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.264140 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3343d19e-07d3-4de8-954a-f7e31aa8279f","Type":"ContainerStarted","Data":"c936fa8962854c1d8665ba026930a8b325915da3444b99cba8eeb7debd2bb042"} Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.264271 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.298952 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.536803277 podStartE2EDuration="22.298928407s" podCreationTimestamp="2026-03-12 18:21:20 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.144925212 +0000 UTC m=+1131.513551545" lastFinishedPulling="2026-03-12 18:21:40.907050342 +0000 UTC m=+1141.275676675" observedRunningTime="2026-03-12 18:21:42.298627567 +0000 UTC m=+1142.667253900" watchObservedRunningTime="2026-03-12 18:21:42.298928407 +0000 UTC m=+1142.667554740" Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.327921 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.651873875 podStartE2EDuration="20.327903948s" podCreationTimestamp="2026-03-12 18:21:22 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.075922266 +0000 UTC m=+1131.444548609" lastFinishedPulling="2026-03-12 18:21:41.751952349 +0000 UTC m=+1142.120578682" observedRunningTime="2026-03-12 18:21:42.324647527 +0000 UTC m=+1142.693273860" watchObservedRunningTime="2026-03-12 18:21:42.327903948 +0000 UTC m=+1142.696530281" Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.331034 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-htjb7"] Mar 12 18:21:42 crc kubenswrapper[4926]: W0312 18:21:42.377170 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7919b4ef_74fc_472e_9a5e_04216cc51ae5.slice/crio-4cd40efc4978929f55aec279245d0d1733c0429e93e1c880bdcbc1f4f5fc9094 WatchSource:0}: Error finding container 4cd40efc4978929f55aec279245d0d1733c0429e93e1c880bdcbc1f4f5fc9094: Status 404 returned error can't find the container with id 4cd40efc4978929f55aec279245d0d1733c0429e93e1c880bdcbc1f4f5fc9094 Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.500900 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" path="/var/lib/kubelet/pods/5241fdef-63ed-416f-9bf0-c004597cc099/volumes" Mar 12 18:21:42 crc kubenswrapper[4926]: I0312 18:21:42.501471 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" path="/var/lib/kubelet/pods/6e12f792-0fbe-4996-a6ec-272e5108dd33/volumes" Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.274396 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06f09c04-6c8d-4c47-a0a5-59def6ebbf94","Type":"ContainerStarted","Data":"cfa97fbf75b6c2e852f5d66a227450ea6c8da1b0641deb3da6fb3c35ba9a0f12"} Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.278493 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9c04aaec-485d-492f-8c24-e6860d9c78f7","Type":"ContainerStarted","Data":"68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5"} Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.280770 4926 generic.go:334] "Generic (PLEG): container finished" podID="30dfa384-92a5-49cf-9793-60478855264f" containerID="88c96a549cb85f5f7b5e96465e3531e8640e2842335f5511d9adb4f48dd39fe6" exitCode=0 Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.280823 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6znbv" event={"ID":"30dfa384-92a5-49cf-9793-60478855264f","Type":"ContainerDied","Data":"88c96a549cb85f5f7b5e96465e3531e8640e2842335f5511d9adb4f48dd39fe6"} Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.282204 4926 generic.go:334] "Generic (PLEG): container finished" podID="10777956-1f18-4bd0-9790-aae020334e2c" containerID="07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586" exitCode=0 Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.282257 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" event={"ID":"10777956-1f18-4bd0-9790-aae020334e2c","Type":"ContainerDied","Data":"07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586"} Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.282278 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" event={"ID":"10777956-1f18-4bd0-9790-aae020334e2c","Type":"ContainerStarted","Data":"1fc31c043cd931246845186f093dded25733b7a182735715b314acd79d3a77b3"} Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.284061 4926 generic.go:334] "Generic (PLEG): container finished" podID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerID="d971505f58d662cfcbae50ecf97fffad61988b03d55fb8228578b1f86349cea3" exitCode=0 Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.284882 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" event={"ID":"7919b4ef-74fc-472e-9a5e-04216cc51ae5","Type":"ContainerDied","Data":"d971505f58d662cfcbae50ecf97fffad61988b03d55fb8228578b1f86349cea3"} Mar 12 18:21:43 crc kubenswrapper[4926]: I0312 18:21:43.284906 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" event={"ID":"7919b4ef-74fc-472e-9a5e-04216cc51ae5","Type":"ContainerStarted","Data":"4cd40efc4978929f55aec279245d0d1733c0429e93e1c880bdcbc1f4f5fc9094"} Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.299944 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" event={"ID":"7919b4ef-74fc-472e-9a5e-04216cc51ae5","Type":"ContainerStarted","Data":"bc88a073a21581ac4854225ad8d64ba64f0d8303fd8d746c072ccb3cc21dc625"} Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.300456 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.303526 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6znbv" event={"ID":"30dfa384-92a5-49cf-9793-60478855264f","Type":"ContainerStarted","Data":"ab5ddbfb0988a75cf57e9b5e666e014d1a4bc30b2f8e2f1f1fae6d36d374f8a0"} Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.303603 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6znbv" event={"ID":"30dfa384-92a5-49cf-9793-60478855264f","Type":"ContainerStarted","Data":"8014dffc7c55d0d12ba9b5c567fd6c20c5f5e73dfb706ab9e919e48b133e0887"} Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.303894 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.315075 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" event={"ID":"10777956-1f18-4bd0-9790-aae020334e2c","Type":"ContainerStarted","Data":"40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d"} Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.322243 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" podStartSLOduration=10.322229426 podStartE2EDuration="10.322229426s" podCreationTimestamp="2026-03-12 18:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:21:44.320413339 +0000 UTC m=+1144.689039672" watchObservedRunningTime="2026-03-12 18:21:44.322229426 +0000 UTC m=+1144.690855759" Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.341547 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" podStartSLOduration=11.341523406 podStartE2EDuration="11.341523406s" podCreationTimestamp="2026-03-12 18:21:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:21:44.33909138 +0000 UTC m=+1144.707717723" watchObservedRunningTime="2026-03-12 18:21:44.341523406 +0000 UTC m=+1144.710149749" Mar 12 18:21:44 crc kubenswrapper[4926]: I0312 18:21:44.367453 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6znbv" podStartSLOduration=9.528671949 podStartE2EDuration="19.367424432s" podCreationTimestamp="2026-03-12 18:21:25 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.878579328 +0000 UTC m=+1132.247205661" lastFinishedPulling="2026-03-12 18:21:41.717331801 +0000 UTC m=+1142.085958144" observedRunningTime="2026-03-12 18:21:44.360483965 +0000 UTC m=+1144.729110318" watchObservedRunningTime="2026-03-12 18:21:44.367424432 +0000 UTC m=+1144.736050755" Mar 12 18:21:45 crc kubenswrapper[4926]: I0312 18:21:45.322813 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:21:45 crc kubenswrapper[4926]: I0312 18:21:45.322881 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:46 crc kubenswrapper[4926]: E0312 18:21:46.022290 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94d38c16_a6c9_44ed_a49e_398dc34b92ce.slice/crio-fc66a1cd90ab8ffc8ce1b57a7c53af70932339daf452369540b2057b973e9dcc.scope\": RecentStats: unable to find data in memory cache]" Mar 12 18:21:46 crc kubenswrapper[4926]: I0312 18:21:46.335966 4926 generic.go:334] "Generic (PLEG): container finished" podID="94d38c16-a6c9-44ed-a49e-398dc34b92ce" containerID="fc66a1cd90ab8ffc8ce1b57a7c53af70932339daf452369540b2057b973e9dcc" exitCode=0 Mar 12 18:21:46 crc kubenswrapper[4926]: I0312 18:21:46.336039 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"94d38c16-a6c9-44ed-a49e-398dc34b92ce","Type":"ContainerDied","Data":"fc66a1cd90ab8ffc8ce1b57a7c53af70932339daf452369540b2057b973e9dcc"} Mar 12 18:21:46 crc kubenswrapper[4926]: I0312 18:21:46.338309 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ee00086-3c8a-4f3a-a5d5-9590715a8b95" containerID="cfea673c23f1afedb464a97e3d967b55a20c7c92aa61908df2cb8c1bc70bc1ba" exitCode=0 Mar 12 18:21:46 crc kubenswrapper[4926]: I0312 18:21:46.339062 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1ee00086-3c8a-4f3a-a5d5-9590715a8b95","Type":"ContainerDied","Data":"cfea673c23f1afedb464a97e3d967b55a20c7c92aa61908df2cb8c1bc70bc1ba"} Mar 12 18:21:49 crc kubenswrapper[4926]: I0312 18:21:49.289626 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:49 crc kubenswrapper[4926]: I0312 18:21:49.462577 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:21:49 crc kubenswrapper[4926]: I0312 18:21:49.555453 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6csg8"] Mar 12 18:21:49 crc kubenswrapper[4926]: I0312 18:21:49.556025 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" podUID="10777956-1f18-4bd0-9790-aae020334e2c" containerName="dnsmasq-dns" containerID="cri-o://40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d" gracePeriod=10 Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.035210 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.135906 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-config\") pod \"10777956-1f18-4bd0-9790-aae020334e2c\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.136046 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28bxt\" (UniqueName: \"kubernetes.io/projected/10777956-1f18-4bd0-9790-aae020334e2c-kube-api-access-28bxt\") pod \"10777956-1f18-4bd0-9790-aae020334e2c\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.136086 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-ovsdbserver-nb\") pod \"10777956-1f18-4bd0-9790-aae020334e2c\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.136108 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-dns-svc\") pod \"10777956-1f18-4bd0-9790-aae020334e2c\" (UID: \"10777956-1f18-4bd0-9790-aae020334e2c\") " Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.146228 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10777956-1f18-4bd0-9790-aae020334e2c-kube-api-access-28bxt" (OuterVolumeSpecName: "kube-api-access-28bxt") pod "10777956-1f18-4bd0-9790-aae020334e2c" (UID: "10777956-1f18-4bd0-9790-aae020334e2c"). InnerVolumeSpecName "kube-api-access-28bxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.172342 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "10777956-1f18-4bd0-9790-aae020334e2c" (UID: "10777956-1f18-4bd0-9790-aae020334e2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.174416 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "10777956-1f18-4bd0-9790-aae020334e2c" (UID: "10777956-1f18-4bd0-9790-aae020334e2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.187050 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-config" (OuterVolumeSpecName: "config") pod "10777956-1f18-4bd0-9790-aae020334e2c" (UID: "10777956-1f18-4bd0-9790-aae020334e2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.237957 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.237998 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28bxt\" (UniqueName: \"kubernetes.io/projected/10777956-1f18-4bd0-9790-aae020334e2c-kube-api-access-28bxt\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.238010 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.238018 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/10777956-1f18-4bd0-9790-aae020334e2c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.389202 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mszqk" event={"ID":"b4a45633-8ac7-497f-a8d7-2b7a3dad35bc","Type":"ContainerStarted","Data":"1ca41fb588e43dfa39c2e3e61a6ab9cc617716eec43113b0c1873a7e2ca9e13a"} Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.392691 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"1ee00086-3c8a-4f3a-a5d5-9590715a8b95","Type":"ContainerStarted","Data":"499f0c68d499ee45ab69e4ac637942df91eb78121e1d774bbabb9618a9b25fff"} Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.396080 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"94d38c16-a6c9-44ed-a49e-398dc34b92ce","Type":"ContainerStarted","Data":"ea2fbdc081b0d7772aa49fd0d249576d8ae27495d6fc4b3dc4b82f68f67285ba"} Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.399068 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c3288572-a9dc-4f96-8535-b05f6f22855b","Type":"ContainerStarted","Data":"4281dca9007224e780bd7011e967ed1045d9d28f90c61c7afe89ae71e702c2a1"} Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.401352 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"8926ccac-b553-4e37-bbcb-96e3b00c1cab","Type":"ContainerStarted","Data":"4cd57bc1175117be13d5d5343e5843a30c935b22b6f485dc1a31a363cd921e85"} Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.403289 4926 generic.go:334] "Generic (PLEG): container finished" podID="10777956-1f18-4bd0-9790-aae020334e2c" containerID="40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d" exitCode=0 Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.403327 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" event={"ID":"10777956-1f18-4bd0-9790-aae020334e2c","Type":"ContainerDied","Data":"40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d"} Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.403348 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" event={"ID":"10777956-1f18-4bd0-9790-aae020334e2c","Type":"ContainerDied","Data":"1fc31c043cd931246845186f093dded25733b7a182735715b314acd79d3a77b3"} Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.403365 4926 scope.go:117] "RemoveContainer" containerID="40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.403492 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-6csg8" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.409417 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-mszqk" podStartSLOduration=10.068310143 podStartE2EDuration="17.40939882s" podCreationTimestamp="2026-03-12 18:21:33 +0000 UTC" firstStartedPulling="2026-03-12 18:21:42.146159943 +0000 UTC m=+1142.514786276" lastFinishedPulling="2026-03-12 18:21:49.48724862 +0000 UTC m=+1149.855874953" observedRunningTime="2026-03-12 18:21:50.408075729 +0000 UTC m=+1150.776702082" watchObservedRunningTime="2026-03-12 18:21:50.40939882 +0000 UTC m=+1150.778025163" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.441907 4926 scope.go:117] "RemoveContainer" containerID="07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.454878 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.909324721 podStartE2EDuration="33.454850354s" podCreationTimestamp="2026-03-12 18:21:17 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.111142431 +0000 UTC m=+1131.479768764" lastFinishedPulling="2026-03-12 18:21:41.656668064 +0000 UTC m=+1142.025294397" observedRunningTime="2026-03-12 18:21:50.445935947 +0000 UTC m=+1150.814562310" watchObservedRunningTime="2026-03-12 18:21:50.454850354 +0000 UTC m=+1150.823476697" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.474094 4926 scope.go:117] "RemoveContainer" containerID="40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d" Mar 12 18:21:50 crc kubenswrapper[4926]: E0312 18:21:50.474813 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d\": container with ID starting with 40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d not found: ID does not exist" containerID="40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.474843 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d"} err="failed to get container status \"40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d\": rpc error: code = NotFound desc = could not find container \"40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d\": container with ID starting with 40485e5bf2b0729b4775af9e0b049a41636e31956c43cb5591bee9a26f2eef0d not found: ID does not exist" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.474862 4926 scope.go:117] "RemoveContainer" containerID="07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586" Mar 12 18:21:50 crc kubenswrapper[4926]: E0312 18:21:50.475140 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586\": container with ID starting with 07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586 not found: ID does not exist" containerID="07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.475160 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586"} err="failed to get container status \"07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586\": rpc error: code = NotFound desc = could not find container \"07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586\": container with ID starting with 07e4167e25a44dc3413aab2abedb803e49deb6627acac90612e2aefe927ed586 not found: ID does not exist" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.498810 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.720888814 podStartE2EDuration="22.498782421s" podCreationTimestamp="2026-03-12 18:21:28 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.702529921 +0000 UTC m=+1132.071156254" lastFinishedPulling="2026-03-12 18:21:49.480423528 +0000 UTC m=+1149.849049861" observedRunningTime="2026-03-12 18:21:50.475045512 +0000 UTC m=+1150.843671845" watchObservedRunningTime="2026-03-12 18:21:50.498782421 +0000 UTC m=+1150.867408764" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.535369 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.996014239 podStartE2EDuration="32.535345289s" podCreationTimestamp="2026-03-12 18:21:18 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.223234849 +0000 UTC m=+1131.591861182" lastFinishedPulling="2026-03-12 18:21:41.762565899 +0000 UTC m=+1142.131192232" observedRunningTime="2026-03-12 18:21:50.516907134 +0000 UTC m=+1150.885533477" watchObservedRunningTime="2026-03-12 18:21:50.535345289 +0000 UTC m=+1150.903971622" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.559646 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.630391 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.526860908 podStartE2EDuration="26.630376115s" podCreationTimestamp="2026-03-12 18:21:24 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.380280165 +0000 UTC m=+1131.748906498" lastFinishedPulling="2026-03-12 18:21:49.483795372 +0000 UTC m=+1149.852421705" observedRunningTime="2026-03-12 18:21:50.594826069 +0000 UTC m=+1150.963452422" watchObservedRunningTime="2026-03-12 18:21:50.630376115 +0000 UTC m=+1150.999002448" Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.657782 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6csg8"] Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.664599 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-6csg8"] Mar 12 18:21:50 crc kubenswrapper[4926]: I0312 18:21:50.972838 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:51 crc kubenswrapper[4926]: I0312 18:21:51.010091 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:51 crc kubenswrapper[4926]: I0312 18:21:51.274540 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:51 crc kubenswrapper[4926]: I0312 18:21:51.411902 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:51 crc kubenswrapper[4926]: I0312 18:21:51.461287 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.499605 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10777956-1f18-4bd0-9790-aae020334e2c" path="/var/lib/kubelet/pods/10777956-1f18-4bd0-9790-aae020334e2c/volumes" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692172 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-7vm4l"] Mar 12 18:21:52 crc kubenswrapper[4926]: E0312 18:21:52.692544 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10777956-1f18-4bd0-9790-aae020334e2c" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692565 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="10777956-1f18-4bd0-9790-aae020334e2c" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: E0312 18:21:52.692595 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692601 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: E0312 18:21:52.692617 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerName="init" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692623 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerName="init" Mar 12 18:21:52 crc kubenswrapper[4926]: E0312 18:21:52.692639 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692645 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: E0312 18:21:52.692660 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" containerName="init" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692667 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" containerName="init" Mar 12 18:21:52 crc kubenswrapper[4926]: E0312 18:21:52.692684 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10777956-1f18-4bd0-9790-aae020334e2c" containerName="init" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692690 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="10777956-1f18-4bd0-9790-aae020334e2c" containerName="init" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692835 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e12f792-0fbe-4996-a6ec-272e5108dd33" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692852 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="10777956-1f18-4bd0-9790-aae020334e2c" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.692865 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5241fdef-63ed-416f-9bf0-c004597cc099" containerName="dnsmasq-dns" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.700497 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.718513 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.739536 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7vm4l"] Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.784163 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-dns-svc\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.784263 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-config\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.784314 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv46v\" (UniqueName: \"kubernetes.io/projected/f569f4cb-b487-41d8-bab4-5c2d7aba2219-kube-api-access-vv46v\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.784345 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.784703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.886915 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv46v\" (UniqueName: \"kubernetes.io/projected/f569f4cb-b487-41d8-bab4-5c2d7aba2219-kube-api-access-vv46v\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.886992 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.887067 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.887093 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-dns-svc\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.887134 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-config\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.887847 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-config\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.888651 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.889268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.889812 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-dns-svc\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:52 crc kubenswrapper[4926]: I0312 18:21:52.913207 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv46v\" (UniqueName: \"kubernetes.io/projected/f569f4cb-b487-41d8-bab4-5c2d7aba2219-kube-api-access-vv46v\") pod \"dnsmasq-dns-698758b865-7vm4l\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.018641 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.279024 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.333362 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.471402 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.573503 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7vm4l"] Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.636633 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.638006 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.646047 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-pvx7j" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.646236 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.646524 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.647257 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.664151 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.799259 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-scripts\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.799329 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.799460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-config\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.799493 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.799525 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.799599 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.799627 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gmzl\" (UniqueName: \"kubernetes.io/projected/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-kube-api-access-6gmzl\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.863573 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.869414 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.871930 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.872085 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.872755 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.876320 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-lnfrb" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.907584 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.909937 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-config\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.909999 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.910049 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.910101 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.910133 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gmzl\" (UniqueName: \"kubernetes.io/projected/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-kube-api-access-6gmzl\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.910150 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-scripts\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.910914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-scripts\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.913291 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-config\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.913591 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.929395 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.932216 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.933730 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.950787 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.953701 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gmzl\" (UniqueName: \"kubernetes.io/projected/3ffbc2f0-cbec-43d4-9907-a95af037ae1b-kube-api-access-6gmzl\") pod \"ovn-northd-0\" (UID: \"3ffbc2f0-cbec-43d4-9907-a95af037ae1b\") " pod="openstack/ovn-northd-0" Mar 12 18:21:53 crc kubenswrapper[4926]: I0312 18:21:53.975971 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.012267 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57853681-32de-4475-9c7d-3f9708fe7d91-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.012325 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts7nq\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-kube-api-access-ts7nq\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.012348 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/57853681-32de-4475-9c7d-3f9708fe7d91-lock\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.012379 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/57853681-32de-4475-9c7d-3f9708fe7d91-cache\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.012403 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.012460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.114302 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/57853681-32de-4475-9c7d-3f9708fe7d91-cache\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.114702 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.114734 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.114796 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57853681-32de-4475-9c7d-3f9708fe7d91-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.114832 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts7nq\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-kube-api-access-ts7nq\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.114850 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/57853681-32de-4475-9c7d-3f9708fe7d91-lock\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.115451 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/57853681-32de-4475-9c7d-3f9708fe7d91-lock\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.115673 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/57853681-32de-4475-9c7d-3f9708fe7d91-cache\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.115908 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: E0312 18:21:54.117997 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 18:21:54 crc kubenswrapper[4926]: E0312 18:21:54.118036 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 18:21:54 crc kubenswrapper[4926]: E0312 18:21:54.118193 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift podName:57853681-32de-4475-9c7d-3f9708fe7d91 nodeName:}" failed. No retries permitted until 2026-03-12 18:21:54.618083775 +0000 UTC m=+1154.986710158 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift") pod "swift-storage-0" (UID: "57853681-32de-4475-9c7d-3f9708fe7d91") : configmap "swift-ring-files" not found Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.124146 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57853681-32de-4475-9c7d-3f9708fe7d91-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.138031 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts7nq\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-kube-api-access-ts7nq\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.147866 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.433260 4926 generic.go:334] "Generic (PLEG): container finished" podID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerID="50524e634e7b1a90622cd86f4a3bdca7cea0ea5d1a412cda03a7f2f2cce56ce2" exitCode=0 Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.433467 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7vm4l" event={"ID":"f569f4cb-b487-41d8-bab4-5c2d7aba2219","Type":"ContainerDied","Data":"50524e634e7b1a90622cd86f4a3bdca7cea0ea5d1a412cda03a7f2f2cce56ce2"} Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.433518 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7vm4l" event={"ID":"f569f4cb-b487-41d8-bab4-5c2d7aba2219","Type":"ContainerStarted","Data":"1a2ca21647a28c29a2f5caa87bf2be645c7eb58039451905ac1f8540e674ede0"} Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.501048 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 12 18:21:54 crc kubenswrapper[4926]: W0312 18:21:54.609006 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ffbc2f0_cbec_43d4_9907_a95af037ae1b.slice/crio-256178c2d03d737a1f12f6ff6236c2beeb58150128480f3343ee00ab2765e9b1 WatchSource:0}: Error finding container 256178c2d03d737a1f12f6ff6236c2beeb58150128480f3343ee00ab2765e9b1: Status 404 returned error can't find the container with id 256178c2d03d737a1f12f6ff6236c2beeb58150128480f3343ee00ab2765e9b1 Mar 12 18:21:54 crc kubenswrapper[4926]: I0312 18:21:54.621685 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:54 crc kubenswrapper[4926]: E0312 18:21:54.622154 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 18:21:54 crc kubenswrapper[4926]: E0312 18:21:54.623228 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 18:21:54 crc kubenswrapper[4926]: E0312 18:21:54.623284 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift podName:57853681-32de-4475-9c7d-3f9708fe7d91 nodeName:}" failed. No retries permitted until 2026-03-12 18:21:55.623266093 +0000 UTC m=+1155.991892446 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift") pod "swift-storage-0" (UID: "57853681-32de-4475-9c7d-3f9708fe7d91") : configmap "swift-ring-files" not found Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.444203 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7vm4l" event={"ID":"f569f4cb-b487-41d8-bab4-5c2d7aba2219","Type":"ContainerStarted","Data":"633289debf3f6b1a2f62e3d378805f919d6393fc63c3e3559b1339530737acbe"} Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.444662 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.445548 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ffbc2f0-cbec-43d4-9907-a95af037ae1b","Type":"ContainerStarted","Data":"256178c2d03d737a1f12f6ff6236c2beeb58150128480f3343ee00ab2765e9b1"} Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.448029 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr" event={"ID":"8f19fbb7-ea3b-437a-a634-498e6a593ef6","Type":"ContainerStarted","Data":"29225e5590e6cbe2eb3df76009321f8119b0d09e042cdfc6354903e1c59fcfc0"} Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.448376 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-sfwpr" Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.495963 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-sfwpr" podStartSLOduration=7.094639749 podStartE2EDuration="30.495946653s" podCreationTimestamp="2026-03-12 18:21:25 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.249857957 +0000 UTC m=+1131.618484290" lastFinishedPulling="2026-03-12 18:21:54.651164861 +0000 UTC m=+1155.019791194" observedRunningTime="2026-03-12 18:21:55.489373608 +0000 UTC m=+1155.857999941" watchObservedRunningTime="2026-03-12 18:21:55.495946653 +0000 UTC m=+1155.864572996" Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.499549 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-7vm4l" podStartSLOduration=3.499544755 podStartE2EDuration="3.499544755s" podCreationTimestamp="2026-03-12 18:21:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:21:55.471838133 +0000 UTC m=+1155.840464466" watchObservedRunningTime="2026-03-12 18:21:55.499544755 +0000 UTC m=+1155.868171088" Mar 12 18:21:55 crc kubenswrapper[4926]: I0312 18:21:55.638002 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:55 crc kubenswrapper[4926]: E0312 18:21:55.638246 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 18:21:55 crc kubenswrapper[4926]: E0312 18:21:55.638268 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 18:21:55 crc kubenswrapper[4926]: E0312 18:21:55.638324 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift podName:57853681-32de-4475-9c7d-3f9708fe7d91 nodeName:}" failed. No retries permitted until 2026-03-12 18:21:57.638305022 +0000 UTC m=+1158.006931355 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift") pod "swift-storage-0" (UID: "57853681-32de-4475-9c7d-3f9708fe7d91") : configmap "swift-ring-files" not found Mar 12 18:21:56 crc kubenswrapper[4926]: I0312 18:21:56.456586 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ffbc2f0-cbec-43d4-9907-a95af037ae1b","Type":"ContainerStarted","Data":"1cb27375fd5a2f047a3e9c961b0ae084ca462a4c2adacc72740c9e181bac727c"} Mar 12 18:21:56 crc kubenswrapper[4926]: I0312 18:21:56.456880 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3ffbc2f0-cbec-43d4-9907-a95af037ae1b","Type":"ContainerStarted","Data":"98ef7afd04dcaf2ad4757cd1e70abbc89963107563e73d273da801e0cb8d6fcc"} Mar 12 18:21:56 crc kubenswrapper[4926]: I0312 18:21:56.457287 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 12 18:21:56 crc kubenswrapper[4926]: I0312 18:21:56.495955 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.44122957 podStartE2EDuration="3.495933645s" podCreationTimestamp="2026-03-12 18:21:53 +0000 UTC" firstStartedPulling="2026-03-12 18:21:54.612120145 +0000 UTC m=+1154.980746498" lastFinishedPulling="2026-03-12 18:21:55.66682424 +0000 UTC m=+1156.035450573" observedRunningTime="2026-03-12 18:21:56.481841687 +0000 UTC m=+1156.850468040" watchObservedRunningTime="2026-03-12 18:21:56.495933645 +0000 UTC m=+1156.864559978" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.680782 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:21:57 crc kubenswrapper[4926]: E0312 18:21:57.680964 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 18:21:57 crc kubenswrapper[4926]: E0312 18:21:57.681271 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 18:21:57 crc kubenswrapper[4926]: E0312 18:21:57.681332 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift podName:57853681-32de-4475-9c7d-3f9708fe7d91 nodeName:}" failed. No retries permitted until 2026-03-12 18:22:01.681308884 +0000 UTC m=+1162.049935297 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift") pod "swift-storage-0" (UID: "57853681-32de-4475-9c7d-3f9708fe7d91") : configmap "swift-ring-files" not found Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.780925 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bn6gf"] Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.782551 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.784858 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.785942 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.791571 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.793732 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bn6gf"] Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.884896 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-ring-data-devices\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.884972 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-etc-swift\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.885029 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bg7n\" (UniqueName: \"kubernetes.io/projected/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-kube-api-access-6bg7n\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.885132 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-scripts\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.885171 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-combined-ca-bundle\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.885260 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-dispersionconf\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.885346 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-swiftconf\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.987595 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-scripts\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.987671 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-combined-ca-bundle\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.987741 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-dispersionconf\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.987823 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-swiftconf\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.987895 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-ring-data-devices\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.987930 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-etc-swift\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.987969 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bg7n\" (UniqueName: \"kubernetes.io/projected/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-kube-api-access-6bg7n\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.988604 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-etc-swift\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.988840 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-scripts\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.988857 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-ring-data-devices\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.994379 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-dispersionconf\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.994570 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-swiftconf\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:57 crc kubenswrapper[4926]: I0312 18:21:57.996719 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-combined-ca-bundle\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:58 crc kubenswrapper[4926]: I0312 18:21:58.019415 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bg7n\" (UniqueName: \"kubernetes.io/projected/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-kube-api-access-6bg7n\") pod \"swift-ring-rebalance-bn6gf\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:58 crc kubenswrapper[4926]: I0312 18:21:58.100008 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:21:58 crc kubenswrapper[4926]: I0312 18:21:58.661907 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 12 18:21:58 crc kubenswrapper[4926]: I0312 18:21:58.662225 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 12 18:21:58 crc kubenswrapper[4926]: I0312 18:21:58.670397 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bn6gf"] Mar 12 18:21:58 crc kubenswrapper[4926]: I0312 18:21:58.750183 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 12 18:21:59 crc kubenswrapper[4926]: I0312 18:21:59.480832 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bn6gf" event={"ID":"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f","Type":"ContainerStarted","Data":"beae81a2d2cc5a26e427094e9b94a43594531b558df0f2e26a7a5160b7c22dda"} Mar 12 18:21:59 crc kubenswrapper[4926]: I0312 18:21:59.549141 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.049495 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.049528 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.130061 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555662-wmwgq"] Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.131208 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555662-wmwgq" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.133534 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.134471 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.134605 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.136505 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555662-wmwgq"] Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.139093 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.229225 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9pnj\" (UniqueName: \"kubernetes.io/projected/3def1951-b6f9-4621-8428-b3e169e34279-kube-api-access-l9pnj\") pod \"auto-csr-approver-29555662-wmwgq\" (UID: \"3def1951-b6f9-4621-8428-b3e169e34279\") " pod="openshift-infra/auto-csr-approver-29555662-wmwgq" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.330285 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9pnj\" (UniqueName: \"kubernetes.io/projected/3def1951-b6f9-4621-8428-b3e169e34279-kube-api-access-l9pnj\") pod \"auto-csr-approver-29555662-wmwgq\" (UID: \"3def1951-b6f9-4621-8428-b3e169e34279\") " pod="openshift-infra/auto-csr-approver-29555662-wmwgq" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.349041 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9pnj\" (UniqueName: \"kubernetes.io/projected/3def1951-b6f9-4621-8428-b3e169e34279-kube-api-access-l9pnj\") pod \"auto-csr-approver-29555662-wmwgq\" (UID: \"3def1951-b6f9-4621-8428-b3e169e34279\") " pod="openshift-infra/auto-csr-approver-29555662-wmwgq" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.453728 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555662-wmwgq" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.589156 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.851252 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-15a0-account-create-update-b4wdv"] Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.860039 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-15a0-account-create-update-b4wdv"] Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.860126 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.862095 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.901170 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-kslsb"] Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.902307 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kslsb" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.916229 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kslsb"] Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.941909 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrl49\" (UniqueName: \"kubernetes.io/projected/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-kube-api-access-jrl49\") pod \"glance-15a0-account-create-update-b4wdv\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:00 crc kubenswrapper[4926]: I0312 18:22:00.942104 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-operator-scripts\") pod \"glance-15a0-account-create-update-b4wdv\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.043386 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-operator-scripts\") pod \"glance-15a0-account-create-update-b4wdv\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.043845 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb8a8892-82ee-4502-b76d-ca289485809b-operator-scripts\") pod \"glance-db-create-kslsb\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " pod="openstack/glance-db-create-kslsb" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.044104 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2gfv\" (UniqueName: \"kubernetes.io/projected/eb8a8892-82ee-4502-b76d-ca289485809b-kube-api-access-v2gfv\") pod \"glance-db-create-kslsb\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " pod="openstack/glance-db-create-kslsb" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.044377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrl49\" (UniqueName: \"kubernetes.io/projected/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-kube-api-access-jrl49\") pod \"glance-15a0-account-create-update-b4wdv\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.044184 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-operator-scripts\") pod \"glance-15a0-account-create-update-b4wdv\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.065278 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrl49\" (UniqueName: \"kubernetes.io/projected/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-kube-api-access-jrl49\") pod \"glance-15a0-account-create-update-b4wdv\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.145773 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb8a8892-82ee-4502-b76d-ca289485809b-operator-scripts\") pod \"glance-db-create-kslsb\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " pod="openstack/glance-db-create-kslsb" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.145850 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2gfv\" (UniqueName: \"kubernetes.io/projected/eb8a8892-82ee-4502-b76d-ca289485809b-kube-api-access-v2gfv\") pod \"glance-db-create-kslsb\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " pod="openstack/glance-db-create-kslsb" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.147022 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb8a8892-82ee-4502-b76d-ca289485809b-operator-scripts\") pod \"glance-db-create-kslsb\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " pod="openstack/glance-db-create-kslsb" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.163096 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2gfv\" (UniqueName: \"kubernetes.io/projected/eb8a8892-82ee-4502-b76d-ca289485809b-kube-api-access-v2gfv\") pod \"glance-db-create-kslsb\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " pod="openstack/glance-db-create-kslsb" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.177933 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.228669 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kslsb" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.487992 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gk626"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.489390 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.496233 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gk626"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.591665 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c596-account-create-update-kw2ql"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.592986 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.595425 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.605350 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c596-account-create-update-kw2ql"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.653130 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-operator-scripts\") pod \"keystone-db-create-gk626\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.654590 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k276h\" (UniqueName: \"kubernetes.io/projected/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-kube-api-access-k276h\") pod \"keystone-db-create-gk626\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.734203 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-87jsw"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.735344 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87jsw" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.752796 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-87jsw"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.756400 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k276h\" (UniqueName: \"kubernetes.io/projected/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-kube-api-access-k276h\") pod \"keystone-db-create-gk626\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.756474 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.756500 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-operator-scripts\") pod \"keystone-db-create-gk626\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.756571 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e559975-5aca-457e-8c50-465552595381-operator-scripts\") pod \"keystone-c596-account-create-update-kw2ql\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.756601 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjz72\" (UniqueName: \"kubernetes.io/projected/1e559975-5aca-457e-8c50-465552595381-kube-api-access-mjz72\") pod \"keystone-c596-account-create-update-kw2ql\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: E0312 18:22:01.756864 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 18:22:01 crc kubenswrapper[4926]: E0312 18:22:01.756894 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 18:22:01 crc kubenswrapper[4926]: E0312 18:22:01.756979 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift podName:57853681-32de-4475-9c7d-3f9708fe7d91 nodeName:}" failed. No retries permitted until 2026-03-12 18:22:09.756961996 +0000 UTC m=+1170.125588329 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift") pod "swift-storage-0" (UID: "57853681-32de-4475-9c7d-3f9708fe7d91") : configmap "swift-ring-files" not found Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.757426 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-operator-scripts\") pod \"keystone-db-create-gk626\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.804343 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k276h\" (UniqueName: \"kubernetes.io/projected/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-kube-api-access-k276h\") pod \"keystone-db-create-gk626\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.809641 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b886-account-create-update-sdqf6"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.811254 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gk626" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.814756 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.817871 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.840601 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b886-account-create-update-sdqf6"] Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.878673 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26s4h\" (UniqueName: \"kubernetes.io/projected/ce8493fd-3e35-41fb-8daa-febd2238ce1b-kube-api-access-26s4h\") pod \"placement-db-create-87jsw\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " pod="openstack/placement-db-create-87jsw" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.878903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e559975-5aca-457e-8c50-465552595381-operator-scripts\") pod \"keystone-c596-account-create-update-kw2ql\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.879047 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjz72\" (UniqueName: \"kubernetes.io/projected/1e559975-5aca-457e-8c50-465552595381-kube-api-access-mjz72\") pod \"keystone-c596-account-create-update-kw2ql\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.879391 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce8493fd-3e35-41fb-8daa-febd2238ce1b-operator-scripts\") pod \"placement-db-create-87jsw\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " pod="openstack/placement-db-create-87jsw" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.879644 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e559975-5aca-457e-8c50-465552595381-operator-scripts\") pod \"keystone-c596-account-create-update-kw2ql\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.900913 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjz72\" (UniqueName: \"kubernetes.io/projected/1e559975-5aca-457e-8c50-465552595381-kube-api-access-mjz72\") pod \"keystone-c596-account-create-update-kw2ql\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.912413 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.980744 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f98d642-e5f2-44de-9259-15b5eed6b80c-operator-scripts\") pod \"placement-b886-account-create-update-sdqf6\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.980797 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8brwx\" (UniqueName: \"kubernetes.io/projected/3f98d642-e5f2-44de-9259-15b5eed6b80c-kube-api-access-8brwx\") pod \"placement-b886-account-create-update-sdqf6\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.980967 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce8493fd-3e35-41fb-8daa-febd2238ce1b-operator-scripts\") pod \"placement-db-create-87jsw\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " pod="openstack/placement-db-create-87jsw" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.981154 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26s4h\" (UniqueName: \"kubernetes.io/projected/ce8493fd-3e35-41fb-8daa-febd2238ce1b-kube-api-access-26s4h\") pod \"placement-db-create-87jsw\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " pod="openstack/placement-db-create-87jsw" Mar 12 18:22:01 crc kubenswrapper[4926]: I0312 18:22:01.981789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce8493fd-3e35-41fb-8daa-febd2238ce1b-operator-scripts\") pod \"placement-db-create-87jsw\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " pod="openstack/placement-db-create-87jsw" Mar 12 18:22:02 crc kubenswrapper[4926]: I0312 18:22:02.011411 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26s4h\" (UniqueName: \"kubernetes.io/projected/ce8493fd-3e35-41fb-8daa-febd2238ce1b-kube-api-access-26s4h\") pod \"placement-db-create-87jsw\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " pod="openstack/placement-db-create-87jsw" Mar 12 18:22:02 crc kubenswrapper[4926]: I0312 18:22:02.057118 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87jsw" Mar 12 18:22:02 crc kubenswrapper[4926]: I0312 18:22:02.086032 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f98d642-e5f2-44de-9259-15b5eed6b80c-operator-scripts\") pod \"placement-b886-account-create-update-sdqf6\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:02 crc kubenswrapper[4926]: I0312 18:22:02.086095 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8brwx\" (UniqueName: \"kubernetes.io/projected/3f98d642-e5f2-44de-9259-15b5eed6b80c-kube-api-access-8brwx\") pod \"placement-b886-account-create-update-sdqf6\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:02 crc kubenswrapper[4926]: I0312 18:22:02.087863 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f98d642-e5f2-44de-9259-15b5eed6b80c-operator-scripts\") pod \"placement-b886-account-create-update-sdqf6\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:02 crc kubenswrapper[4926]: I0312 18:22:02.105686 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8brwx\" (UniqueName: \"kubernetes.io/projected/3f98d642-e5f2-44de-9259-15b5eed6b80c-kube-api-access-8brwx\") pod \"placement-b886-account-create-update-sdqf6\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:02 crc kubenswrapper[4926]: I0312 18:22:02.136079 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.023641 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.129870 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-htjb7"] Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.130133 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" podUID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerName="dnsmasq-dns" containerID="cri-o://bc88a073a21581ac4854225ad8d64ba64f0d8303fd8d746c072ccb3cc21dc625" gracePeriod=10 Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.466028 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-kslsb"] Mar 12 18:22:03 crc kubenswrapper[4926]: W0312 18:22:03.543903 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3def1951_b6f9_4621_8428_b3e169e34279.slice/crio-9995b1f578722c4da205a824f7fc383db43c0b559444cc089533ad24be578f7c WatchSource:0}: Error finding container 9995b1f578722c4da205a824f7fc383db43c0b559444cc089533ad24be578f7c: Status 404 returned error can't find the container with id 9995b1f578722c4da205a824f7fc383db43c0b559444cc089533ad24be578f7c Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.546393 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555662-wmwgq"] Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.548201 4926 generic.go:334] "Generic (PLEG): container finished" podID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerID="bc88a073a21581ac4854225ad8d64ba64f0d8303fd8d746c072ccb3cc21dc625" exitCode=0 Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.548312 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" event={"ID":"7919b4ef-74fc-472e-9a5e-04216cc51ae5","Type":"ContainerDied","Data":"bc88a073a21581ac4854225ad8d64ba64f0d8303fd8d746c072ccb3cc21dc625"} Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.550343 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bn6gf" event={"ID":"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f","Type":"ContainerStarted","Data":"251cc81f683a75e3d21cb79796d4e3d3d467bddd70539c194beb7a011504559b"} Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.552932 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kslsb" event={"ID":"eb8a8892-82ee-4502-b76d-ca289485809b","Type":"ContainerStarted","Data":"00b76c7be6d0fdb000f31c1e38cf50ca0b2d1f46c57eb45e800949b818b564a5"} Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.560565 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c596-account-create-update-kw2ql"] Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.579086 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bn6gf" podStartSLOduration=2.513716254 podStartE2EDuration="6.579068886s" podCreationTimestamp="2026-03-12 18:21:57 +0000 UTC" firstStartedPulling="2026-03-12 18:21:58.679219611 +0000 UTC m=+1159.047845944" lastFinishedPulling="2026-03-12 18:22:02.744572243 +0000 UTC m=+1163.113198576" observedRunningTime="2026-03-12 18:22:03.566070501 +0000 UTC m=+1163.934696834" watchObservedRunningTime="2026-03-12 18:22:03.579068886 +0000 UTC m=+1163.947695219" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.652553 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.760967 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-15a0-account-create-update-b4wdv"] Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.772518 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-87jsw"] Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.782270 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b886-account-create-update-sdqf6"] Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.786694 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gk626"] Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.828228 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-config\") pod \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.828427 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-sb\") pod \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.828477 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7c2g\" (UniqueName: \"kubernetes.io/projected/7919b4ef-74fc-472e-9a5e-04216cc51ae5-kube-api-access-l7c2g\") pod \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.828503 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-dns-svc\") pod \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.828544 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-nb\") pod \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\" (UID: \"7919b4ef-74fc-472e-9a5e-04216cc51ae5\") " Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.845408 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7919b4ef-74fc-472e-9a5e-04216cc51ae5-kube-api-access-l7c2g" (OuterVolumeSpecName: "kube-api-access-l7c2g") pod "7919b4ef-74fc-472e-9a5e-04216cc51ae5" (UID: "7919b4ef-74fc-472e-9a5e-04216cc51ae5"). InnerVolumeSpecName "kube-api-access-l7c2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.888703 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7919b4ef-74fc-472e-9a5e-04216cc51ae5" (UID: "7919b4ef-74fc-472e-9a5e-04216cc51ae5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.931763 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.931836 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7c2g\" (UniqueName: \"kubernetes.io/projected/7919b4ef-74fc-472e-9a5e-04216cc51ae5-kube-api-access-l7c2g\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.932931 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7919b4ef-74fc-472e-9a5e-04216cc51ae5" (UID: "7919b4ef-74fc-472e-9a5e-04216cc51ae5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.933057 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7919b4ef-74fc-472e-9a5e-04216cc51ae5" (UID: "7919b4ef-74fc-472e-9a5e-04216cc51ae5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:03 crc kubenswrapper[4926]: I0312 18:22:03.943975 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-config" (OuterVolumeSpecName: "config") pod "7919b4ef-74fc-472e-9a5e-04216cc51ae5" (UID: "7919b4ef-74fc-472e-9a5e-04216cc51ae5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.033172 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.033202 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.033216 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7919b4ef-74fc-472e-9a5e-04216cc51ae5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.565571 4926 generic.go:334] "Generic (PLEG): container finished" podID="ce8493fd-3e35-41fb-8daa-febd2238ce1b" containerID="0477440cc1caee314ece47874cba75dc234908b5f0302ff1737f87d526dec4a4" exitCode=0 Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.565733 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-87jsw" event={"ID":"ce8493fd-3e35-41fb-8daa-febd2238ce1b","Type":"ContainerDied","Data":"0477440cc1caee314ece47874cba75dc234908b5f0302ff1737f87d526dec4a4"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.565826 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-87jsw" event={"ID":"ce8493fd-3e35-41fb-8daa-febd2238ce1b","Type":"ContainerStarted","Data":"2fffaad6d5a06f655ff4c8d188a160c681cfc6e69dc4ef7ffcfa90f4d6f269b4"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.568090 4926 generic.go:334] "Generic (PLEG): container finished" podID="1e559975-5aca-457e-8c50-465552595381" containerID="ed6d1abff4f7da34f442e943b17a81aeae55a22054d604db8a2f0c813a7bb890" exitCode=0 Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.568135 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c596-account-create-update-kw2ql" event={"ID":"1e559975-5aca-457e-8c50-465552595381","Type":"ContainerDied","Data":"ed6d1abff4f7da34f442e943b17a81aeae55a22054d604db8a2f0c813a7bb890"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.568152 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c596-account-create-update-kw2ql" event={"ID":"1e559975-5aca-457e-8c50-465552595381","Type":"ContainerStarted","Data":"12f3735d737a130d7a9095b30d8233ba764f6de01a5c6fe30de6758919c90b92"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.570553 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b886-account-create-update-sdqf6" event={"ID":"3f98d642-e5f2-44de-9259-15b5eed6b80c","Type":"ContainerStarted","Data":"2ba173648181598e4717e6ee2a92ecb7ee6351d447d5a38a4302f5af077322b5"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.570582 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b886-account-create-update-sdqf6" event={"ID":"3f98d642-e5f2-44de-9259-15b5eed6b80c","Type":"ContainerStarted","Data":"2b00f6a8874e175c9a39535d556a6fbc2f2a3ee6efb123f9a4087969dcab6675"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.574825 4926 generic.go:334] "Generic (PLEG): container finished" podID="eb8a8892-82ee-4502-b76d-ca289485809b" containerID="d5aba66a32fd03fd9f7cad091c15aa0c8e52be10f9b3914b0869cd5e59465cde" exitCode=0 Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.574884 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kslsb" event={"ID":"eb8a8892-82ee-4502-b76d-ca289485809b","Type":"ContainerDied","Data":"d5aba66a32fd03fd9f7cad091c15aa0c8e52be10f9b3914b0869cd5e59465cde"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.581094 4926 generic.go:334] "Generic (PLEG): container finished" podID="51bf320d-e5c9-43c6-baf7-e1f2f9ee3313" containerID="97dffad6ad390315746d80bb46b2150bb89f6d5359806877446913c809356199" exitCode=0 Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.581159 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gk626" event={"ID":"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313","Type":"ContainerDied","Data":"97dffad6ad390315746d80bb46b2150bb89f6d5359806877446913c809356199"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.581188 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gk626" event={"ID":"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313","Type":"ContainerStarted","Data":"2dfd9c4d7dc7e12afe66ccf426b7bea55f65eaade1defa6244ce76ada3b5ba5c"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.584833 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555662-wmwgq" event={"ID":"3def1951-b6f9-4621-8428-b3e169e34279","Type":"ContainerStarted","Data":"9995b1f578722c4da205a824f7fc383db43c0b559444cc089533ad24be578f7c"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.586952 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496" containerID="7093c3070454c81ffcfc913032818ad12b03f50c370515a66252d6a038735742" exitCode=0 Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.587009 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-15a0-account-create-update-b4wdv" event={"ID":"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496","Type":"ContainerDied","Data":"7093c3070454c81ffcfc913032818ad12b03f50c370515a66252d6a038735742"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.587025 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-15a0-account-create-update-b4wdv" event={"ID":"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496","Type":"ContainerStarted","Data":"bf0bf6219a3827f9978a4d657d9f0c8902ab0058726263be468bd1610eca879f"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.591033 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.591565 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-htjb7" event={"ID":"7919b4ef-74fc-472e-9a5e-04216cc51ae5","Type":"ContainerDied","Data":"4cd40efc4978929f55aec279245d0d1733c0429e93e1c880bdcbc1f4f5fc9094"} Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.591651 4926 scope.go:117] "RemoveContainer" containerID="bc88a073a21581ac4854225ad8d64ba64f0d8303fd8d746c072ccb3cc21dc625" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.603220 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-b886-account-create-update-sdqf6" podStartSLOduration=3.603199459 podStartE2EDuration="3.603199459s" podCreationTimestamp="2026-03-12 18:22:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:04.594194848 +0000 UTC m=+1164.962821191" watchObservedRunningTime="2026-03-12 18:22:04.603199459 +0000 UTC m=+1164.971825792" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.639167 4926 scope.go:117] "RemoveContainer" containerID="d971505f58d662cfcbae50ecf97fffad61988b03d55fb8228578b1f86349cea3" Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.692117 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-htjb7"] Mar 12 18:22:04 crc kubenswrapper[4926]: I0312 18:22:04.698318 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-htjb7"] Mar 12 18:22:05 crc kubenswrapper[4926]: I0312 18:22:05.602534 4926 generic.go:334] "Generic (PLEG): container finished" podID="3f98d642-e5f2-44de-9259-15b5eed6b80c" containerID="2ba173648181598e4717e6ee2a92ecb7ee6351d447d5a38a4302f5af077322b5" exitCode=0 Mar 12 18:22:05 crc kubenswrapper[4926]: I0312 18:22:05.602605 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b886-account-create-update-sdqf6" event={"ID":"3f98d642-e5f2-44de-9259-15b5eed6b80c","Type":"ContainerDied","Data":"2ba173648181598e4717e6ee2a92ecb7ee6351d447d5a38a4302f5af077322b5"} Mar 12 18:22:05 crc kubenswrapper[4926]: I0312 18:22:05.604157 4926 generic.go:334] "Generic (PLEG): container finished" podID="3def1951-b6f9-4621-8428-b3e169e34279" containerID="6e11e8ecbd719cce51e25090959e2c12962023683d089714c24c4c379fbeaa75" exitCode=0 Mar 12 18:22:05 crc kubenswrapper[4926]: I0312 18:22:05.604607 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555662-wmwgq" event={"ID":"3def1951-b6f9-4621-8428-b3e169e34279","Type":"ContainerDied","Data":"6e11e8ecbd719cce51e25090959e2c12962023683d089714c24c4c379fbeaa75"} Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.122006 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gk626" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.266647 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.277642 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.291063 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-operator-scripts\") pod \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.291166 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k276h\" (UniqueName: \"kubernetes.io/projected/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-kube-api-access-k276h\") pod \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\" (UID: \"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.291895 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51bf320d-e5c9-43c6-baf7-e1f2f9ee3313" (UID: "51bf320d-e5c9-43c6-baf7-e1f2f9ee3313"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.305344 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-kube-api-access-k276h" (OuterVolumeSpecName: "kube-api-access-k276h") pod "51bf320d-e5c9-43c6-baf7-e1f2f9ee3313" (UID: "51bf320d-e5c9-43c6-baf7-e1f2f9ee3313"). InnerVolumeSpecName "kube-api-access-k276h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.311432 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kslsb" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.347317 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87jsw" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.392338 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e559975-5aca-457e-8c50-465552595381-operator-scripts\") pod \"1e559975-5aca-457e-8c50-465552595381\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.392481 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrl49\" (UniqueName: \"kubernetes.io/projected/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-kube-api-access-jrl49\") pod \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.392527 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjz72\" (UniqueName: \"kubernetes.io/projected/1e559975-5aca-457e-8c50-465552595381-kube-api-access-mjz72\") pod \"1e559975-5aca-457e-8c50-465552595381\" (UID: \"1e559975-5aca-457e-8c50-465552595381\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.392630 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-operator-scripts\") pod \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\" (UID: \"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.392795 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e559975-5aca-457e-8c50-465552595381-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1e559975-5aca-457e-8c50-465552595381" (UID: "1e559975-5aca-457e-8c50-465552595381"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.393176 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k276h\" (UniqueName: \"kubernetes.io/projected/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-kube-api-access-k276h\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.393201 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1e559975-5aca-457e-8c50-465552595381-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.393213 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.393682 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496" (UID: "2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.399991 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-kube-api-access-jrl49" (OuterVolumeSpecName: "kube-api-access-jrl49") pod "2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496" (UID: "2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496"). InnerVolumeSpecName "kube-api-access-jrl49". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.400102 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e559975-5aca-457e-8c50-465552595381-kube-api-access-mjz72" (OuterVolumeSpecName: "kube-api-access-mjz72") pod "1e559975-5aca-457e-8c50-465552595381" (UID: "1e559975-5aca-457e-8c50-465552595381"). InnerVolumeSpecName "kube-api-access-mjz72". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.494818 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26s4h\" (UniqueName: \"kubernetes.io/projected/ce8493fd-3e35-41fb-8daa-febd2238ce1b-kube-api-access-26s4h\") pod \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.494919 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb8a8892-82ee-4502-b76d-ca289485809b-operator-scripts\") pod \"eb8a8892-82ee-4502-b76d-ca289485809b\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.495175 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce8493fd-3e35-41fb-8daa-febd2238ce1b-operator-scripts\") pod \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\" (UID: \"ce8493fd-3e35-41fb-8daa-febd2238ce1b\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.495220 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2gfv\" (UniqueName: \"kubernetes.io/projected/eb8a8892-82ee-4502-b76d-ca289485809b-kube-api-access-v2gfv\") pod \"eb8a8892-82ee-4502-b76d-ca289485809b\" (UID: \"eb8a8892-82ee-4502-b76d-ca289485809b\") " Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.495682 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce8493fd-3e35-41fb-8daa-febd2238ce1b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce8493fd-3e35-41fb-8daa-febd2238ce1b" (UID: "ce8493fd-3e35-41fb-8daa-febd2238ce1b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.495912 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrl49\" (UniqueName: \"kubernetes.io/projected/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-kube-api-access-jrl49\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.495932 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjz72\" (UniqueName: \"kubernetes.io/projected/1e559975-5aca-457e-8c50-465552595381-kube-api-access-mjz72\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.495944 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce8493fd-3e35-41fb-8daa-febd2238ce1b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.495959 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.496581 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb8a8892-82ee-4502-b76d-ca289485809b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eb8a8892-82ee-4502-b76d-ca289485809b" (UID: "eb8a8892-82ee-4502-b76d-ca289485809b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.498490 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce8493fd-3e35-41fb-8daa-febd2238ce1b-kube-api-access-26s4h" (OuterVolumeSpecName: "kube-api-access-26s4h") pod "ce8493fd-3e35-41fb-8daa-febd2238ce1b" (UID: "ce8493fd-3e35-41fb-8daa-febd2238ce1b"). InnerVolumeSpecName "kube-api-access-26s4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.498559 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb8a8892-82ee-4502-b76d-ca289485809b-kube-api-access-v2gfv" (OuterVolumeSpecName: "kube-api-access-v2gfv") pod "eb8a8892-82ee-4502-b76d-ca289485809b" (UID: "eb8a8892-82ee-4502-b76d-ca289485809b"). InnerVolumeSpecName "kube-api-access-v2gfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.504475 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" path="/var/lib/kubelet/pods/7919b4ef-74fc-472e-9a5e-04216cc51ae5/volumes" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.597091 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2gfv\" (UniqueName: \"kubernetes.io/projected/eb8a8892-82ee-4502-b76d-ca289485809b-kube-api-access-v2gfv\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.597135 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26s4h\" (UniqueName: \"kubernetes.io/projected/ce8493fd-3e35-41fb-8daa-febd2238ce1b-kube-api-access-26s4h\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.597150 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb8a8892-82ee-4502-b76d-ca289485809b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.612669 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-87jsw" event={"ID":"ce8493fd-3e35-41fb-8daa-febd2238ce1b","Type":"ContainerDied","Data":"2fffaad6d5a06f655ff4c8d188a160c681cfc6e69dc4ef7ffcfa90f4d6f269b4"} Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.612692 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-87jsw" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.612714 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fffaad6d5a06f655ff4c8d188a160c681cfc6e69dc4ef7ffcfa90f4d6f269b4" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.616257 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c596-account-create-update-kw2ql" event={"ID":"1e559975-5aca-457e-8c50-465552595381","Type":"ContainerDied","Data":"12f3735d737a130d7a9095b30d8233ba764f6de01a5c6fe30de6758919c90b92"} Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.616426 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12f3735d737a130d7a9095b30d8233ba764f6de01a5c6fe30de6758919c90b92" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.616352 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c596-account-create-update-kw2ql" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.617686 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-kslsb" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.617703 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-kslsb" event={"ID":"eb8a8892-82ee-4502-b76d-ca289485809b","Type":"ContainerDied","Data":"00b76c7be6d0fdb000f31c1e38cf50ca0b2d1f46c57eb45e800949b818b564a5"} Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.617747 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00b76c7be6d0fdb000f31c1e38cf50ca0b2d1f46c57eb45e800949b818b564a5" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.619975 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gk626" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.620382 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gk626" event={"ID":"51bf320d-e5c9-43c6-baf7-e1f2f9ee3313","Type":"ContainerDied","Data":"2dfd9c4d7dc7e12afe66ccf426b7bea55f65eaade1defa6244ce76ada3b5ba5c"} Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.620742 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dfd9c4d7dc7e12afe66ccf426b7bea55f65eaade1defa6244ce76ada3b5ba5c" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.621512 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-15a0-account-create-update-b4wdv" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.622515 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-15a0-account-create-update-b4wdv" event={"ID":"2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496","Type":"ContainerDied","Data":"bf0bf6219a3827f9978a4d657d9f0c8902ab0058726263be468bd1610eca879f"} Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.622825 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf0bf6219a3827f9978a4d657d9f0c8902ab0058726263be468bd1610eca879f" Mar 12 18:22:06 crc kubenswrapper[4926]: I0312 18:22:06.966554 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555662-wmwgq" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.104408 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9pnj\" (UniqueName: \"kubernetes.io/projected/3def1951-b6f9-4621-8428-b3e169e34279-kube-api-access-l9pnj\") pod \"3def1951-b6f9-4621-8428-b3e169e34279\" (UID: \"3def1951-b6f9-4621-8428-b3e169e34279\") " Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.108632 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3def1951-b6f9-4621-8428-b3e169e34279-kube-api-access-l9pnj" (OuterVolumeSpecName: "kube-api-access-l9pnj") pod "3def1951-b6f9-4621-8428-b3e169e34279" (UID: "3def1951-b6f9-4621-8428-b3e169e34279"). InnerVolumeSpecName "kube-api-access-l9pnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.114049 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.206092 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9pnj\" (UniqueName: \"kubernetes.io/projected/3def1951-b6f9-4621-8428-b3e169e34279-kube-api-access-l9pnj\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.307352 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8brwx\" (UniqueName: \"kubernetes.io/projected/3f98d642-e5f2-44de-9259-15b5eed6b80c-kube-api-access-8brwx\") pod \"3f98d642-e5f2-44de-9259-15b5eed6b80c\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.307743 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f98d642-e5f2-44de-9259-15b5eed6b80c-operator-scripts\") pod \"3f98d642-e5f2-44de-9259-15b5eed6b80c\" (UID: \"3f98d642-e5f2-44de-9259-15b5eed6b80c\") " Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.308395 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f98d642-e5f2-44de-9259-15b5eed6b80c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f98d642-e5f2-44de-9259-15b5eed6b80c" (UID: "3f98d642-e5f2-44de-9259-15b5eed6b80c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.311906 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f98d642-e5f2-44de-9259-15b5eed6b80c-kube-api-access-8brwx" (OuterVolumeSpecName: "kube-api-access-8brwx") pod "3f98d642-e5f2-44de-9259-15b5eed6b80c" (UID: "3f98d642-e5f2-44de-9259-15b5eed6b80c"). InnerVolumeSpecName "kube-api-access-8brwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.318916 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-bgd7w"] Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319423 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerName="dnsmasq-dns" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319479 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerName="dnsmasq-dns" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319508 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb8a8892-82ee-4502-b76d-ca289485809b" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319520 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb8a8892-82ee-4502-b76d-ca289485809b" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319545 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f98d642-e5f2-44de-9259-15b5eed6b80c" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319558 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f98d642-e5f2-44de-9259-15b5eed6b80c" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319600 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e559975-5aca-457e-8c50-465552595381" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319612 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e559975-5aca-457e-8c50-465552595381" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319632 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce8493fd-3e35-41fb-8daa-febd2238ce1b" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319644 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce8493fd-3e35-41fb-8daa-febd2238ce1b" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319664 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerName="init" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319676 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerName="init" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319694 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3def1951-b6f9-4621-8428-b3e169e34279" containerName="oc" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319706 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3def1951-b6f9-4621-8428-b3e169e34279" containerName="oc" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319733 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319745 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: E0312 18:22:07.319763 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51bf320d-e5c9-43c6-baf7-e1f2f9ee3313" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.319776 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="51bf320d-e5c9-43c6-baf7-e1f2f9ee3313" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320064 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7919b4ef-74fc-472e-9a5e-04216cc51ae5" containerName="dnsmasq-dns" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320091 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3def1951-b6f9-4621-8428-b3e169e34279" containerName="oc" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320112 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f98d642-e5f2-44de-9259-15b5eed6b80c" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320132 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e559975-5aca-457e-8c50-465552595381" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320146 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496" containerName="mariadb-account-create-update" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320160 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="51bf320d-e5c9-43c6-baf7-e1f2f9ee3313" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320182 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce8493fd-3e35-41fb-8daa-febd2238ce1b" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.320203 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb8a8892-82ee-4502-b76d-ca289485809b" containerName="mariadb-database-create" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.321008 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.327497 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.335166 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bgd7w"] Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.410320 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8brwx\" (UniqueName: \"kubernetes.io/projected/3f98d642-e5f2-44de-9259-15b5eed6b80c-kube-api-access-8brwx\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.410365 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f98d642-e5f2-44de-9259-15b5eed6b80c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.512235 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x52cf\" (UniqueName: \"kubernetes.io/projected/88fe4c28-3abc-43de-8686-83c78b0633f8-kube-api-access-x52cf\") pod \"root-account-create-update-bgd7w\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.512302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88fe4c28-3abc-43de-8686-83c78b0633f8-operator-scripts\") pod \"root-account-create-update-bgd7w\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.614729 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x52cf\" (UniqueName: \"kubernetes.io/projected/88fe4c28-3abc-43de-8686-83c78b0633f8-kube-api-access-x52cf\") pod \"root-account-create-update-bgd7w\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.614859 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88fe4c28-3abc-43de-8686-83c78b0633f8-operator-scripts\") pod \"root-account-create-update-bgd7w\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.617766 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88fe4c28-3abc-43de-8686-83c78b0633f8-operator-scripts\") pod \"root-account-create-update-bgd7w\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.659385 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555662-wmwgq" event={"ID":"3def1951-b6f9-4621-8428-b3e169e34279","Type":"ContainerDied","Data":"9995b1f578722c4da205a824f7fc383db43c0b559444cc089533ad24be578f7c"} Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.659473 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9995b1f578722c4da205a824f7fc383db43c0b559444cc089533ad24be578f7c" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.659550 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555662-wmwgq" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.675168 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x52cf\" (UniqueName: \"kubernetes.io/projected/88fe4c28-3abc-43de-8686-83c78b0633f8-kube-api-access-x52cf\") pod \"root-account-create-update-bgd7w\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.677375 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b886-account-create-update-sdqf6" event={"ID":"3f98d642-e5f2-44de-9259-15b5eed6b80c","Type":"ContainerDied","Data":"2b00f6a8874e175c9a39535d556a6fbc2f2a3ee6efb123f9a4087969dcab6675"} Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.677431 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b00f6a8874e175c9a39535d556a6fbc2f2a3ee6efb123f9a4087969dcab6675" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.677527 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b886-account-create-update-sdqf6" Mar 12 18:22:07 crc kubenswrapper[4926]: I0312 18:22:07.682169 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:08 crc kubenswrapper[4926]: I0312 18:22:08.063295 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555656-zlrgw"] Mar 12 18:22:08 crc kubenswrapper[4926]: I0312 18:22:08.071423 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555656-zlrgw"] Mar 12 18:22:08 crc kubenswrapper[4926]: I0312 18:22:08.171941 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-bgd7w"] Mar 12 18:22:08 crc kubenswrapper[4926]: I0312 18:22:08.507914 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b042fb81-959e-48c0-8a9e-87bafcad2fe3" path="/var/lib/kubelet/pods/b042fb81-959e-48c0-8a9e-87bafcad2fe3/volumes" Mar 12 18:22:08 crc kubenswrapper[4926]: I0312 18:22:08.688146 4926 generic.go:334] "Generic (PLEG): container finished" podID="88fe4c28-3abc-43de-8686-83c78b0633f8" containerID="bf4088b7010b072d07c97127da27603d49f76d54fb965fcf30cdce478bed5715" exitCode=0 Mar 12 18:22:08 crc kubenswrapper[4926]: I0312 18:22:08.688187 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bgd7w" event={"ID":"88fe4c28-3abc-43de-8686-83c78b0633f8","Type":"ContainerDied","Data":"bf4088b7010b072d07c97127da27603d49f76d54fb965fcf30cdce478bed5715"} Mar 12 18:22:08 crc kubenswrapper[4926]: I0312 18:22:08.688212 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bgd7w" event={"ID":"88fe4c28-3abc-43de-8686-83c78b0633f8","Type":"ContainerStarted","Data":"ff6a1dc4d902432ca748504cde2a8c61dd124de912bd1cf55d1e3457e983d254"} Mar 12 18:22:09 crc kubenswrapper[4926]: I0312 18:22:09.853537 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:22:09 crc kubenswrapper[4926]: E0312 18:22:09.853842 4926 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 12 18:22:09 crc kubenswrapper[4926]: E0312 18:22:09.853874 4926 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 12 18:22:09 crc kubenswrapper[4926]: E0312 18:22:09.854036 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift podName:57853681-32de-4475-9c7d-3f9708fe7d91 nodeName:}" failed. No retries permitted until 2026-03-12 18:22:25.854009492 +0000 UTC m=+1186.222635825 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift") pod "swift-storage-0" (UID: "57853681-32de-4475-9c7d-3f9708fe7d91") : configmap "swift-ring-files" not found Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.038959 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.159292 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x52cf\" (UniqueName: \"kubernetes.io/projected/88fe4c28-3abc-43de-8686-83c78b0633f8-kube-api-access-x52cf\") pod \"88fe4c28-3abc-43de-8686-83c78b0633f8\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.159647 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88fe4c28-3abc-43de-8686-83c78b0633f8-operator-scripts\") pod \"88fe4c28-3abc-43de-8686-83c78b0633f8\" (UID: \"88fe4c28-3abc-43de-8686-83c78b0633f8\") " Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.160226 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88fe4c28-3abc-43de-8686-83c78b0633f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "88fe4c28-3abc-43de-8686-83c78b0633f8" (UID: "88fe4c28-3abc-43de-8686-83c78b0633f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.166713 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88fe4c28-3abc-43de-8686-83c78b0633f8-kube-api-access-x52cf" (OuterVolumeSpecName: "kube-api-access-x52cf") pod "88fe4c28-3abc-43de-8686-83c78b0633f8" (UID: "88fe4c28-3abc-43de-8686-83c78b0633f8"). InnerVolumeSpecName "kube-api-access-x52cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.261700 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x52cf\" (UniqueName: \"kubernetes.io/projected/88fe4c28-3abc-43de-8686-83c78b0633f8-kube-api-access-x52cf\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.261726 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88fe4c28-3abc-43de-8686-83c78b0633f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.711467 4926 generic.go:334] "Generic (PLEG): container finished" podID="9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" containerID="251cc81f683a75e3d21cb79796d4e3d3d467bddd70539c194beb7a011504559b" exitCode=0 Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.711559 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bn6gf" event={"ID":"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f","Type":"ContainerDied","Data":"251cc81f683a75e3d21cb79796d4e3d3d467bddd70539c194beb7a011504559b"} Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.714143 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-bgd7w" event={"ID":"88fe4c28-3abc-43de-8686-83c78b0633f8","Type":"ContainerDied","Data":"ff6a1dc4d902432ca748504cde2a8c61dd124de912bd1cf55d1e3457e983d254"} Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.714200 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff6a1dc4d902432ca748504cde2a8c61dd124de912bd1cf55d1e3457e983d254" Mar 12 18:22:10 crc kubenswrapper[4926]: I0312 18:22:10.714522 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-bgd7w" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.193914 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-72msf"] Mar 12 18:22:11 crc kubenswrapper[4926]: E0312 18:22:11.194198 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88fe4c28-3abc-43de-8686-83c78b0633f8" containerName="mariadb-account-create-update" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.194210 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="88fe4c28-3abc-43de-8686-83c78b0633f8" containerName="mariadb-account-create-update" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.194357 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="88fe4c28-3abc-43de-8686-83c78b0633f8" containerName="mariadb-account-create-update" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.194848 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.198082 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2zfbp" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.204953 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.207668 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-72msf"] Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.280760 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-db-sync-config-data\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.280829 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-combined-ca-bundle\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.280880 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9kjm\" (UniqueName: \"kubernetes.io/projected/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-kube-api-access-q9kjm\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.280911 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-config-data\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.382829 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9kjm\" (UniqueName: \"kubernetes.io/projected/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-kube-api-access-q9kjm\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.382896 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-config-data\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.383049 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-db-sync-config-data\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.383095 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-combined-ca-bundle\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.386655 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-db-sync-config-data\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.388009 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-combined-ca-bundle\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.392087 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-config-data\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.409016 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9kjm\" (UniqueName: \"kubernetes.io/projected/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-kube-api-access-q9kjm\") pod \"glance-db-sync-72msf\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " pod="openstack/glance-db-sync-72msf" Mar 12 18:22:11 crc kubenswrapper[4926]: I0312 18:22:11.518198 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72msf" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.040688 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.072559 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-72msf"] Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.199887 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bg7n\" (UniqueName: \"kubernetes.io/projected/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-kube-api-access-6bg7n\") pod \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.199977 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-dispersionconf\") pod \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.200069 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-ring-data-devices\") pod \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.200114 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-combined-ca-bundle\") pod \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.200208 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-etc-swift\") pod \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.200255 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-swiftconf\") pod \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.200318 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-scripts\") pod \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\" (UID: \"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f\") " Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.202121 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" (UID: "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.202237 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" (UID: "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.206726 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-kube-api-access-6bg7n" (OuterVolumeSpecName: "kube-api-access-6bg7n") pod "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" (UID: "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f"). InnerVolumeSpecName "kube-api-access-6bg7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.212152 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" (UID: "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.223007 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-scripts" (OuterVolumeSpecName: "scripts") pod "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" (UID: "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.229716 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" (UID: "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.236903 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" (UID: "9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.302533 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.302569 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bg7n\" (UniqueName: \"kubernetes.io/projected/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-kube-api-access-6bg7n\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.302580 4926 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.302590 4926 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.302599 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.302607 4926 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.302618 4926 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.730034 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bn6gf" event={"ID":"9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f","Type":"ContainerDied","Data":"beae81a2d2cc5a26e427094e9b94a43594531b558df0f2e26a7a5160b7c22dda"} Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.730085 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="beae81a2d2cc5a26e427094e9b94a43594531b558df0f2e26a7a5160b7c22dda" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.730094 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bn6gf" Mar 12 18:22:12 crc kubenswrapper[4926]: I0312 18:22:12.730991 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72msf" event={"ID":"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8","Type":"ContainerStarted","Data":"0b8074caa3b0c10eb9a072ab0513f4ac17eb3bec44cc3edb55302dc941d344c0"} Mar 12 18:22:13 crc kubenswrapper[4926]: I0312 18:22:13.709000 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-bgd7w"] Mar 12 18:22:13 crc kubenswrapper[4926]: I0312 18:22:13.715672 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-bgd7w"] Mar 12 18:22:14 crc kubenswrapper[4926]: I0312 18:22:14.047201 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 12 18:22:14 crc kubenswrapper[4926]: I0312 18:22:14.499062 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88fe4c28-3abc-43de-8686-83c78b0633f8" path="/var/lib/kubelet/pods/88fe4c28-3abc-43de-8686-83c78b0633f8/volumes" Mar 12 18:22:15 crc kubenswrapper[4926]: I0312 18:22:15.768569 4926 generic.go:334] "Generic (PLEG): container finished" podID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerID="68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5" exitCode=0 Mar 12 18:22:15 crc kubenswrapper[4926]: I0312 18:22:15.768970 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9c04aaec-485d-492f-8c24-e6860d9c78f7","Type":"ContainerDied","Data":"68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5"} Mar 12 18:22:15 crc kubenswrapper[4926]: I0312 18:22:15.773336 4926 generic.go:334] "Generic (PLEG): container finished" podID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerID="cfa97fbf75b6c2e852f5d66a227450ea6c8da1b0641deb3da6fb3c35ba9a0f12" exitCode=0 Mar 12 18:22:15 crc kubenswrapper[4926]: I0312 18:22:15.773369 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06f09c04-6c8d-4c47-a0a5-59def6ebbf94","Type":"ContainerDied","Data":"cfa97fbf75b6c2e852f5d66a227450ea6c8da1b0641deb3da6fb3c35ba9a0f12"} Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.258026 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.269587 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6znbv" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.547685 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sfwpr-config-m7tgt"] Mar 12 18:22:16 crc kubenswrapper[4926]: E0312 18:22:16.548880 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" containerName="swift-ring-rebalance" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.548904 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" containerName="swift-ring-rebalance" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.549772 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f" containerName="swift-ring-rebalance" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.551371 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfwpr-config-m7tgt"] Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.551637 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.554872 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.702754 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-scripts\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.703027 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9bbf\" (UniqueName: \"kubernetes.io/projected/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-kube-api-access-v9bbf\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.703155 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-additional-scripts\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.703271 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.703393 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run-ovn\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.703634 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-log-ovn\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.806179 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-scripts\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.806257 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9bbf\" (UniqueName: \"kubernetes.io/projected/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-kube-api-access-v9bbf\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.806284 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-additional-scripts\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.806322 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.806362 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run-ovn\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.806406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-log-ovn\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.806722 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-log-ovn\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.808064 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-additional-scripts\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.808397 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.808588 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run-ovn\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.808604 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-scripts\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.809109 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9c04aaec-485d-492f-8c24-e6860d9c78f7","Type":"ContainerStarted","Data":"ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a"} Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.810181 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.819407 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06f09c04-6c8d-4c47-a0a5-59def6ebbf94","Type":"ContainerStarted","Data":"3e4a5dd026300b0565c4a11203b217973524caa9bcc9839e71d49583b866e88e"} Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.820345 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.854432 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9bbf\" (UniqueName: \"kubernetes.io/projected/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-kube-api-access-v9bbf\") pod \"ovn-controller-sfwpr-config-m7tgt\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.862144 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.244234415 podStartE2EDuration="1m0.862124989s" podCreationTimestamp="2026-03-12 18:21:16 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.099916532 +0000 UTC m=+1131.468542865" lastFinishedPulling="2026-03-12 18:21:41.717807106 +0000 UTC m=+1142.086433439" observedRunningTime="2026-03-12 18:22:16.847796143 +0000 UTC m=+1177.216422476" watchObservedRunningTime="2026-03-12 18:22:16.862124989 +0000 UTC m=+1177.230751322" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.879432 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:16 crc kubenswrapper[4926]: I0312 18:22:16.889804 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.307808794 podStartE2EDuration="1m0.889749979s" podCreationTimestamp="2026-03-12 18:21:16 +0000 UTC" firstStartedPulling="2026-03-12 18:21:31.325478209 +0000 UTC m=+1131.694104542" lastFinishedPulling="2026-03-12 18:21:40.907419394 +0000 UTC m=+1141.276045727" observedRunningTime="2026-03-12 18:22:16.885823736 +0000 UTC m=+1177.254450069" watchObservedRunningTime="2026-03-12 18:22:16.889749979 +0000 UTC m=+1177.258376312" Mar 12 18:22:17 crc kubenswrapper[4926]: I0312 18:22:17.430465 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfwpr-config-m7tgt"] Mar 12 18:22:17 crc kubenswrapper[4926]: I0312 18:22:17.826763 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr-config-m7tgt" event={"ID":"eab8aa6f-7146-49ab-bedf-e7dae33f96bf","Type":"ContainerStarted","Data":"f80e900658b9e00858dd34680664f8b822bb4e0eba4897392ac1a1da7732a56c"} Mar 12 18:22:17 crc kubenswrapper[4926]: I0312 18:22:17.827025 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr-config-m7tgt" event={"ID":"eab8aa6f-7146-49ab-bedf-e7dae33f96bf","Type":"ContainerStarted","Data":"63ca14d801a286c7a6026983b605cad8b07e6eedc355463549b31b37de421eb4"} Mar 12 18:22:17 crc kubenswrapper[4926]: I0312 18:22:17.842555 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-sfwpr-config-m7tgt" podStartSLOduration=1.842538472 podStartE2EDuration="1.842538472s" podCreationTimestamp="2026-03-12 18:22:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:17.842175501 +0000 UTC m=+1178.210801834" watchObservedRunningTime="2026-03-12 18:22:17.842538472 +0000 UTC m=+1178.211164805" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.716288 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-jlt78"] Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.717742 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.721895 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.728863 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-jlt78"] Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.839696 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvnd6\" (UniqueName: \"kubernetes.io/projected/349ca4f5-349b-45ab-98a4-844fa00599d0-kube-api-access-tvnd6\") pod \"root-account-create-update-jlt78\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.839747 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ca4f5-349b-45ab-98a4-844fa00599d0-operator-scripts\") pod \"root-account-create-update-jlt78\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.851800 4926 generic.go:334] "Generic (PLEG): container finished" podID="eab8aa6f-7146-49ab-bedf-e7dae33f96bf" containerID="f80e900658b9e00858dd34680664f8b822bb4e0eba4897392ac1a1da7732a56c" exitCode=0 Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.851838 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr-config-m7tgt" event={"ID":"eab8aa6f-7146-49ab-bedf-e7dae33f96bf","Type":"ContainerDied","Data":"f80e900658b9e00858dd34680664f8b822bb4e0eba4897392ac1a1da7732a56c"} Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.941148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvnd6\" (UniqueName: \"kubernetes.io/projected/349ca4f5-349b-45ab-98a4-844fa00599d0-kube-api-access-tvnd6\") pod \"root-account-create-update-jlt78\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.941230 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ca4f5-349b-45ab-98a4-844fa00599d0-operator-scripts\") pod \"root-account-create-update-jlt78\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.942083 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ca4f5-349b-45ab-98a4-844fa00599d0-operator-scripts\") pod \"root-account-create-update-jlt78\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:18 crc kubenswrapper[4926]: I0312 18:22:18.969478 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvnd6\" (UniqueName: \"kubernetes.io/projected/349ca4f5-349b-45ab-98a4-844fa00599d0-kube-api-access-tvnd6\") pod \"root-account-create-update-jlt78\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:19 crc kubenswrapper[4926]: I0312 18:22:19.034129 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:25 crc kubenswrapper[4926]: I0312 18:22:25.867405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:22:26 crc kubenswrapper[4926]: I0312 18:22:25.876342 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/57853681-32de-4475-9c7d-3f9708fe7d91-etc-swift\") pod \"swift-storage-0\" (UID: \"57853681-32de-4475-9c7d-3f9708fe7d91\") " pod="openstack/swift-storage-0" Mar 12 18:22:26 crc kubenswrapper[4926]: I0312 18:22:26.011875 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 12 18:22:26 crc kubenswrapper[4926]: I0312 18:22:26.230131 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-sfwpr" Mar 12 18:22:27 crc kubenswrapper[4926]: I0312 18:22:27.506717 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:22:27 crc kubenswrapper[4926]: I0312 18:22:27.771628 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 12 18:22:27 crc kubenswrapper[4926]: I0312 18:22:27.965498 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:27 crc kubenswrapper[4926]: I0312 18:22:27.976257 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr-config-m7tgt" event={"ID":"eab8aa6f-7146-49ab-bedf-e7dae33f96bf","Type":"ContainerDied","Data":"63ca14d801a286c7a6026983b605cad8b07e6eedc355463549b31b37de421eb4"} Mar 12 18:22:27 crc kubenswrapper[4926]: I0312 18:22:27.976521 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63ca14d801a286c7a6026983b605cad8b07e6eedc355463549b31b37de421eb4" Mar 12 18:22:27 crc kubenswrapper[4926]: I0312 18:22:27.976404 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-m7tgt" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004645 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9bbf\" (UniqueName: \"kubernetes.io/projected/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-kube-api-access-v9bbf\") pod \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run-ovn\") pod \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004797 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-scripts\") pod \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004860 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-log-ovn\") pod \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004890 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run\") pod \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004914 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "eab8aa6f-7146-49ab-bedf-e7dae33f96bf" (UID: "eab8aa6f-7146-49ab-bedf-e7dae33f96bf"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004939 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-additional-scripts\") pod \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\" (UID: \"eab8aa6f-7146-49ab-bedf-e7dae33f96bf\") " Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.004965 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "eab8aa6f-7146-49ab-bedf-e7dae33f96bf" (UID: "eab8aa6f-7146-49ab-bedf-e7dae33f96bf"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.005028 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run" (OuterVolumeSpecName: "var-run") pod "eab8aa6f-7146-49ab-bedf-e7dae33f96bf" (UID: "eab8aa6f-7146-49ab-bedf-e7dae33f96bf"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.005515 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "eab8aa6f-7146-49ab-bedf-e7dae33f96bf" (UID: "eab8aa6f-7146-49ab-bedf-e7dae33f96bf"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.005621 4926 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.005635 4926 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.005646 4926 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-var-run\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.005657 4926 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.005818 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-scripts" (OuterVolumeSpecName: "scripts") pod "eab8aa6f-7146-49ab-bedf-e7dae33f96bf" (UID: "eab8aa6f-7146-49ab-bedf-e7dae33f96bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.013507 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-kube-api-access-v9bbf" (OuterVolumeSpecName: "kube-api-access-v9bbf") pod "eab8aa6f-7146-49ab-bedf-e7dae33f96bf" (UID: "eab8aa6f-7146-49ab-bedf-e7dae33f96bf"). InnerVolumeSpecName "kube-api-access-v9bbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.107523 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9bbf\" (UniqueName: \"kubernetes.io/projected/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-kube-api-access-v9bbf\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.107559 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab8aa6f-7146-49ab-bedf-e7dae33f96bf-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.313769 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-jlt78"] Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.502267 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 12 18:22:28 crc kubenswrapper[4926]: W0312 18:22:28.505300 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57853681_32de_4475_9c7d_3f9708fe7d91.slice/crio-d44e9a8ee5ceb248e04686de962e1c80c3449a6c3bd013213aff0a7fef9a9e4c WatchSource:0}: Error finding container d44e9a8ee5ceb248e04686de962e1c80c3449a6c3bd013213aff0a7fef9a9e4c: Status 404 returned error can't find the container with id d44e9a8ee5ceb248e04686de962e1c80c3449a6c3bd013213aff0a7fef9a9e4c Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.984072 4926 generic.go:334] "Generic (PLEG): container finished" podID="349ca4f5-349b-45ab-98a4-844fa00599d0" containerID="c3e155687eb67c8841a428e422ae982d9a3f9a577c50dd1606a11fd1ad93242c" exitCode=0 Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.984141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-jlt78" event={"ID":"349ca4f5-349b-45ab-98a4-844fa00599d0","Type":"ContainerDied","Data":"c3e155687eb67c8841a428e422ae982d9a3f9a577c50dd1606a11fd1ad93242c"} Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.984164 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-jlt78" event={"ID":"349ca4f5-349b-45ab-98a4-844fa00599d0","Type":"ContainerStarted","Data":"5ee494a44c685214964380e7c1b459b8f01159d69ae5d0dd6ee963a73f27e878"} Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.985369 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72msf" event={"ID":"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8","Type":"ContainerStarted","Data":"f8a8e8e5fcda402042df2295b1c2cbd321aa245d9b7be946af5559093b56242a"} Mar 12 18:22:28 crc kubenswrapper[4926]: I0312 18:22:28.986236 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"d44e9a8ee5ceb248e04686de962e1c80c3449a6c3bd013213aff0a7fef9a9e4c"} Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.031908 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-72msf" podStartSLOduration=2.241839521 podStartE2EDuration="18.031883259s" podCreationTimestamp="2026-03-12 18:22:11 +0000 UTC" firstStartedPulling="2026-03-12 18:22:12.075050894 +0000 UTC m=+1172.443677247" lastFinishedPulling="2026-03-12 18:22:27.865094652 +0000 UTC m=+1188.233720985" observedRunningTime="2026-03-12 18:22:29.025243513 +0000 UTC m=+1189.393869856" watchObservedRunningTime="2026-03-12 18:22:29.031883259 +0000 UTC m=+1189.400509592" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.098576 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-sfwpr-config-m7tgt"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.104225 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-sfwpr-config-m7tgt"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.201504 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-sfwpr-config-55drx"] Mar 12 18:22:29 crc kubenswrapper[4926]: E0312 18:22:29.202263 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab8aa6f-7146-49ab-bedf-e7dae33f96bf" containerName="ovn-config" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.202276 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab8aa6f-7146-49ab-bedf-e7dae33f96bf" containerName="ovn-config" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.202451 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab8aa6f-7146-49ab-bedf-e7dae33f96bf" containerName="ovn-config" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.202961 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.204913 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.208975 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfwpr-config-55drx"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.330779 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-additional-scripts\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.330839 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-scripts\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.330918 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drskx\" (UniqueName: \"kubernetes.io/projected/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-kube-api-access-drskx\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.331014 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-log-ovn\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.331164 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.331297 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run-ovn\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.405465 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-n9jj2"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.406501 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.425837 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-n9jj2"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433164 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-additional-scripts\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433212 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-scripts\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433263 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drskx\" (UniqueName: \"kubernetes.io/projected/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-kube-api-access-drskx\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433283 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-log-ovn\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433321 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433357 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run-ovn\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433642 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run-ovn\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433645 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-log-ovn\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.433661 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.434525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-additional-scripts\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.439354 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-scripts\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.498663 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drskx\" (UniqueName: \"kubernetes.io/projected/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-kube-api-access-drskx\") pod \"ovn-controller-sfwpr-config-55drx\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.509773 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-40f3-account-create-update-6f5kz"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.510886 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.513620 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.519095 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-40f3-account-create-update-6f5kz"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.528102 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.534585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rddff\" (UniqueName: \"kubernetes.io/projected/fa60a9bd-1933-44b8-ac17-be1f72c9da68-kube-api-access-rddff\") pod \"cinder-db-create-n9jj2\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.534626 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa60a9bd-1933-44b8-ac17-be1f72c9da68-operator-scripts\") pod \"cinder-db-create-n9jj2\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.595364 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-n28hs"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.599699 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.608340 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-n28hs"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.635623 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rddff\" (UniqueName: \"kubernetes.io/projected/fa60a9bd-1933-44b8-ac17-be1f72c9da68-kube-api-access-rddff\") pod \"cinder-db-create-n9jj2\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.635664 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa60a9bd-1933-44b8-ac17-be1f72c9da68-operator-scripts\") pod \"cinder-db-create-n9jj2\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.635712 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pb5c\" (UniqueName: \"kubernetes.io/projected/6a1a6049-a924-4e9e-adad-e6ec84732eb9-kube-api-access-9pb5c\") pod \"cinder-40f3-account-create-update-6f5kz\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.635758 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1a6049-a924-4e9e-adad-e6ec84732eb9-operator-scripts\") pod \"cinder-40f3-account-create-update-6f5kz\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.636769 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa60a9bd-1933-44b8-ac17-be1f72c9da68-operator-scripts\") pod \"cinder-db-create-n9jj2\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.672082 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rddff\" (UniqueName: \"kubernetes.io/projected/fa60a9bd-1933-44b8-ac17-be1f72c9da68-kube-api-access-rddff\") pod \"cinder-db-create-n9jj2\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.714986 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-bf50-account-create-update-ckzh6"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.716184 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.722114 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.723571 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.737149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05959443-d099-4653-9736-7745ba1ce331-operator-scripts\") pod \"barbican-db-create-n28hs\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.737238 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pb5c\" (UniqueName: \"kubernetes.io/projected/6a1a6049-a924-4e9e-adad-e6ec84732eb9-kube-api-access-9pb5c\") pod \"cinder-40f3-account-create-update-6f5kz\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.737275 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7tgz\" (UniqueName: \"kubernetes.io/projected/05959443-d099-4653-9736-7745ba1ce331-kube-api-access-w7tgz\") pod \"barbican-db-create-n28hs\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.737314 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1a6049-a924-4e9e-adad-e6ec84732eb9-operator-scripts\") pod \"cinder-40f3-account-create-update-6f5kz\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.738100 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1a6049-a924-4e9e-adad-e6ec84732eb9-operator-scripts\") pod \"cinder-40f3-account-create-update-6f5kz\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.744068 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bf50-account-create-update-ckzh6"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.763288 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pb5c\" (UniqueName: \"kubernetes.io/projected/6a1a6049-a924-4e9e-adad-e6ec84732eb9-kube-api-access-9pb5c\") pod \"cinder-40f3-account-create-update-6f5kz\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.789013 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-4vtkq"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.789985 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.794687 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.794875 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.795428 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.795636 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vrsc7" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.825927 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-qr6zk"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.826857 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.834368 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.838804 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-kube-api-access-8nh4b\") pod \"barbican-bf50-account-create-update-ckzh6\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.838921 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7tgz\" (UniqueName: \"kubernetes.io/projected/05959443-d099-4653-9736-7745ba1ce331-kube-api-access-w7tgz\") pod \"barbican-db-create-n28hs\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.838972 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-operator-scripts\") pod \"barbican-bf50-account-create-update-ckzh6\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.839008 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-config-data\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.839069 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05959443-d099-4653-9736-7745ba1ce331-operator-scripts\") pod \"barbican-db-create-n28hs\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.839101 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8k85\" (UniqueName: \"kubernetes.io/projected/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-kube-api-access-w8k85\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.839120 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-combined-ca-bundle\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.840255 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05959443-d099-4653-9736-7745ba1ce331-operator-scripts\") pod \"barbican-db-create-n28hs\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.858511 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-qr6zk"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.869693 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-2c76-account-create-update-wvgzd"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.870682 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.872512 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.879904 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7tgz\" (UniqueName: \"kubernetes.io/projected/05959443-d099-4653-9736-7745ba1ce331-kube-api-access-w7tgz\") pod \"barbican-db-create-n28hs\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.892496 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2c76-account-create-update-wvgzd"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.907880 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4vtkq"] Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.935512 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940308 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8k85\" (UniqueName: \"kubernetes.io/projected/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-kube-api-access-w8k85\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940353 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-combined-ca-bundle\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940392 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-kube-api-access-8nh4b\") pod \"barbican-bf50-account-create-update-ckzh6\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940427 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jl7q\" (UniqueName: \"kubernetes.io/projected/ea314c69-7524-43d6-9d4f-9fdb16510952-kube-api-access-7jl7q\") pod \"neutron-db-create-qr6zk\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940502 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-operator-scripts\") pod \"barbican-bf50-account-create-update-ckzh6\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940559 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea314c69-7524-43d6-9d4f-9fdb16510952-operator-scripts\") pod \"neutron-db-create-qr6zk\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940585 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-config-data\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940615 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7917312-5922-44df-a838-6b5452e8bb84-operator-scripts\") pod \"neutron-2c76-account-create-update-wvgzd\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.940648 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r6dd\" (UniqueName: \"kubernetes.io/projected/f7917312-5922-44df-a838-6b5452e8bb84-kube-api-access-7r6dd\") pod \"neutron-2c76-account-create-update-wvgzd\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.941307 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-operator-scripts\") pod \"barbican-bf50-account-create-update-ckzh6\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.945731 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-combined-ca-bundle\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.951080 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-config-data\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.964959 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8k85\" (UniqueName: \"kubernetes.io/projected/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-kube-api-access-w8k85\") pod \"keystone-db-sync-4vtkq\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:29 crc kubenswrapper[4926]: I0312 18:22:29.982175 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-kube-api-access-8nh4b\") pod \"barbican-bf50-account-create-update-ckzh6\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.037059 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.042491 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r6dd\" (UniqueName: \"kubernetes.io/projected/f7917312-5922-44df-a838-6b5452e8bb84-kube-api-access-7r6dd\") pod \"neutron-2c76-account-create-update-wvgzd\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.042668 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jl7q\" (UniqueName: \"kubernetes.io/projected/ea314c69-7524-43d6-9d4f-9fdb16510952-kube-api-access-7jl7q\") pod \"neutron-db-create-qr6zk\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.042727 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea314c69-7524-43d6-9d4f-9fdb16510952-operator-scripts\") pod \"neutron-db-create-qr6zk\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.042790 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7917312-5922-44df-a838-6b5452e8bb84-operator-scripts\") pod \"neutron-2c76-account-create-update-wvgzd\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.043677 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7917312-5922-44df-a838-6b5452e8bb84-operator-scripts\") pod \"neutron-2c76-account-create-update-wvgzd\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.043691 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea314c69-7524-43d6-9d4f-9fdb16510952-operator-scripts\") pod \"neutron-db-create-qr6zk\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.070974 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jl7q\" (UniqueName: \"kubernetes.io/projected/ea314c69-7524-43d6-9d4f-9fdb16510952-kube-api-access-7jl7q\") pod \"neutron-db-create-qr6zk\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.070995 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r6dd\" (UniqueName: \"kubernetes.io/projected/f7917312-5922-44df-a838-6b5452e8bb84-kube-api-access-7r6dd\") pod \"neutron-2c76-account-create-update-wvgzd\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.150511 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-sfwpr-config-55drx"] Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.250182 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.266677 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.275991 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.316767 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-40f3-account-create-update-6f5kz"] Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.432909 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-n9jj2"] Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.500793 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab8aa6f-7146-49ab-bedf-e7dae33f96bf" path="/var/lib/kubelet/pods/eab8aa6f-7146-49ab-bedf-e7dae33f96bf/volumes" Mar 12 18:22:30 crc kubenswrapper[4926]: W0312 18:22:30.793903 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a1a6049_a924_4e9e_adad_e6ec84732eb9.slice/crio-7f60838faf0603a5159f5f3ca03ddcba8992db0a0a7cb8b0a847b3f2ddc0a013 WatchSource:0}: Error finding container 7f60838faf0603a5159f5f3ca03ddcba8992db0a0a7cb8b0a847b3f2ddc0a013: Status 404 returned error can't find the container with id 7f60838faf0603a5159f5f3ca03ddcba8992db0a0a7cb8b0a847b3f2ddc0a013 Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.855314 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.958575 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ca4f5-349b-45ab-98a4-844fa00599d0-operator-scripts\") pod \"349ca4f5-349b-45ab-98a4-844fa00599d0\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.958803 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvnd6\" (UniqueName: \"kubernetes.io/projected/349ca4f5-349b-45ab-98a4-844fa00599d0-kube-api-access-tvnd6\") pod \"349ca4f5-349b-45ab-98a4-844fa00599d0\" (UID: \"349ca4f5-349b-45ab-98a4-844fa00599d0\") " Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.962575 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/349ca4f5-349b-45ab-98a4-844fa00599d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "349ca4f5-349b-45ab-98a4-844fa00599d0" (UID: "349ca4f5-349b-45ab-98a4-844fa00599d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:30 crc kubenswrapper[4926]: I0312 18:22:30.968510 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/349ca4f5-349b-45ab-98a4-844fa00599d0-kube-api-access-tvnd6" (OuterVolumeSpecName: "kube-api-access-tvnd6") pod "349ca4f5-349b-45ab-98a4-844fa00599d0" (UID: "349ca4f5-349b-45ab-98a4-844fa00599d0"). InnerVolumeSpecName "kube-api-access-tvnd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.066548 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/349ca4f5-349b-45ab-98a4-844fa00599d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.066875 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvnd6\" (UniqueName: \"kubernetes.io/projected/349ca4f5-349b-45ab-98a4-844fa00599d0-kube-api-access-tvnd6\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.083020 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-jlt78" Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.084263 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-jlt78" event={"ID":"349ca4f5-349b-45ab-98a4-844fa00599d0","Type":"ContainerDied","Data":"5ee494a44c685214964380e7c1b459b8f01159d69ae5d0dd6ee963a73f27e878"} Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.084306 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ee494a44c685214964380e7c1b459b8f01159d69ae5d0dd6ee963a73f27e878" Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.087952 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-40f3-account-create-update-6f5kz" event={"ID":"6a1a6049-a924-4e9e-adad-e6ec84732eb9","Type":"ContainerStarted","Data":"7f60838faf0603a5159f5f3ca03ddcba8992db0a0a7cb8b0a847b3f2ddc0a013"} Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.098682 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-n9jj2" event={"ID":"fa60a9bd-1933-44b8-ac17-be1f72c9da68","Type":"ContainerStarted","Data":"2b292237feadafee95b31e06cfc840f72c515e88b8454c8337e1e74e28cd7576"} Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.133384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr-config-55drx" event={"ID":"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d","Type":"ContainerStarted","Data":"1b64556f5bf261422bd7a946c6906ac510a7650f9110c56570560ce6cb0ceff7"} Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.531119 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-2c76-account-create-update-wvgzd"] Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.538400 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-n28hs"] Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.615644 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bf50-account-create-update-ckzh6"] Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.626821 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-qr6zk"] Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.719903 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4vtkq"] Mar 12 18:22:31 crc kubenswrapper[4926]: W0312 18:22:31.730812 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea314c69_7524_43d6_9d4f_9fdb16510952.slice/crio-c5f3bf41c45721fdd76098163bfbbd5fb8a4d739371fec426b54c95a8d7546f5 WatchSource:0}: Error finding container c5f3bf41c45721fdd76098163bfbbd5fb8a4d739371fec426b54c95a8d7546f5: Status 404 returned error can't find the container with id c5f3bf41c45721fdd76098163bfbbd5fb8a4d739371fec426b54c95a8d7546f5 Mar 12 18:22:31 crc kubenswrapper[4926]: W0312 18:22:31.733915 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e5bf33c_5cb6_4b29_95f4_fcf47183be58.slice/crio-c56db79b42dcfa1fc516e7e174210c09707255662ee7602092c441dc8b95945d WatchSource:0}: Error finding container c56db79b42dcfa1fc516e7e174210c09707255662ee7602092c441dc8b95945d: Status 404 returned error can't find the container with id c56db79b42dcfa1fc516e7e174210c09707255662ee7602092c441dc8b95945d Mar 12 18:22:31 crc kubenswrapper[4926]: I0312 18:22:31.739711 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.141052 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4vtkq" event={"ID":"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d","Type":"ContainerStarted","Data":"7c7e1f1cb85b710df8d9fc682c9ab5a089b0604e2b56bc8fd5036a48bcce3088"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.142844 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2c76-account-create-update-wvgzd" event={"ID":"f7917312-5922-44df-a838-6b5452e8bb84","Type":"ContainerStarted","Data":"1a95a2291b71efd71b4d69e1bff55bb10de0a0cbc11d599a6fbdc537cf2a1c21"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.142876 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2c76-account-create-update-wvgzd" event={"ID":"f7917312-5922-44df-a838-6b5452e8bb84","Type":"ContainerStarted","Data":"63fccee8b612554f7b094c76fecdac4ac35f19b95dc6b9a6784a3936c615d9a5"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.144358 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bf50-account-create-update-ckzh6" event={"ID":"9e5bf33c-5cb6-4b29-95f4-fcf47183be58","Type":"ContainerStarted","Data":"ff84043d3f919db7adaecbb9d5f2f9199eabe47965d6033ec791be724e00a8f2"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.144380 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bf50-account-create-update-ckzh6" event={"ID":"9e5bf33c-5cb6-4b29-95f4-fcf47183be58","Type":"ContainerStarted","Data":"c56db79b42dcfa1fc516e7e174210c09707255662ee7602092c441dc8b95945d"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.146013 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-qr6zk" event={"ID":"ea314c69-7524-43d6-9d4f-9fdb16510952","Type":"ContainerStarted","Data":"7a2cc30c7cfe8ca6acf3b5029da578678f6497bf01186f5ff5a4caf50683498c"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.146056 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-qr6zk" event={"ID":"ea314c69-7524-43d6-9d4f-9fdb16510952","Type":"ContainerStarted","Data":"c5f3bf41c45721fdd76098163bfbbd5fb8a4d739371fec426b54c95a8d7546f5"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.147580 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-n28hs" event={"ID":"05959443-d099-4653-9736-7745ba1ce331","Type":"ContainerStarted","Data":"f7b35eaf50020882859b6b648c035a75d02c47868f155d967f93cf393c62f39e"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.147615 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-n28hs" event={"ID":"05959443-d099-4653-9736-7745ba1ce331","Type":"ContainerStarted","Data":"def54e7f3bca68561150fcadb909db84a99aaf5cfe2458f9499bbf1014d7ccf8"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.149173 4926 generic.go:334] "Generic (PLEG): container finished" podID="fa60a9bd-1933-44b8-ac17-be1f72c9da68" containerID="12643d715a7a25afe43a9b8c6f2f23a51e1343ba5e8cfc3caebf03243d5ac109" exitCode=0 Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.149239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-n9jj2" event={"ID":"fa60a9bd-1933-44b8-ac17-be1f72c9da68","Type":"ContainerDied","Data":"12643d715a7a25afe43a9b8c6f2f23a51e1343ba5e8cfc3caebf03243d5ac109"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.150819 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" containerID="f2d7197b9b3bf88ce380eff3700bdf47487a3209f60cede9380f4a5b74fd2a07" exitCode=0 Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.150858 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr-config-55drx" event={"ID":"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d","Type":"ContainerDied","Data":"f2d7197b9b3bf88ce380eff3700bdf47487a3209f60cede9380f4a5b74fd2a07"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.154076 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"3ed8b79685248ae25cde3e72f762ac91192ada9775d542cdf62b1115d6fa4f4e"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.154106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"872a3c331ab2c6465434c4ee0f90bfa5304ded6c4675c8f9b4a055f0cc249520"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.154116 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"c8d0099df2ef97a7e69d18ad8472352f918822e9cbc54b1cba03b6ac41eaeb75"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.155517 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a1a6049-a924-4e9e-adad-e6ec84732eb9" containerID="c4054fe5394aadede666252204c9b503dc03ddd88691e230f2deda1936fe7bc5" exitCode=0 Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.155557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-40f3-account-create-update-6f5kz" event={"ID":"6a1a6049-a924-4e9e-adad-e6ec84732eb9","Type":"ContainerDied","Data":"c4054fe5394aadede666252204c9b503dc03ddd88691e230f2deda1936fe7bc5"} Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.158992 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-2c76-account-create-update-wvgzd" podStartSLOduration=3.158975574 podStartE2EDuration="3.158975574s" podCreationTimestamp="2026-03-12 18:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:32.1572663 +0000 UTC m=+1192.525892633" watchObservedRunningTime="2026-03-12 18:22:32.158975574 +0000 UTC m=+1192.527601907" Mar 12 18:22:32 crc kubenswrapper[4926]: I0312 18:22:32.192297 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-n28hs" podStartSLOduration=3.19228327 podStartE2EDuration="3.19228327s" podCreationTimestamp="2026-03-12 18:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:32.184318032 +0000 UTC m=+1192.552944365" watchObservedRunningTime="2026-03-12 18:22:32.19228327 +0000 UTC m=+1192.560909603" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.164711 4926 generic.go:334] "Generic (PLEG): container finished" podID="05959443-d099-4653-9736-7745ba1ce331" containerID="f7b35eaf50020882859b6b648c035a75d02c47868f155d967f93cf393c62f39e" exitCode=0 Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.164801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-n28hs" event={"ID":"05959443-d099-4653-9736-7745ba1ce331","Type":"ContainerDied","Data":"f7b35eaf50020882859b6b648c035a75d02c47868f155d967f93cf393c62f39e"} Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.169095 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7917312-5922-44df-a838-6b5452e8bb84" containerID="1a95a2291b71efd71b4d69e1bff55bb10de0a0cbc11d599a6fbdc537cf2a1c21" exitCode=0 Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.169189 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2c76-account-create-update-wvgzd" event={"ID":"f7917312-5922-44df-a838-6b5452e8bb84","Type":"ContainerDied","Data":"1a95a2291b71efd71b4d69e1bff55bb10de0a0cbc11d599a6fbdc537cf2a1c21"} Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.170749 4926 generic.go:334] "Generic (PLEG): container finished" podID="9e5bf33c-5cb6-4b29-95f4-fcf47183be58" containerID="ff84043d3f919db7adaecbb9d5f2f9199eabe47965d6033ec791be724e00a8f2" exitCode=0 Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.170800 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bf50-account-create-update-ckzh6" event={"ID":"9e5bf33c-5cb6-4b29-95f4-fcf47183be58","Type":"ContainerDied","Data":"ff84043d3f919db7adaecbb9d5f2f9199eabe47965d6033ec791be724e00a8f2"} Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.173020 4926 generic.go:334] "Generic (PLEG): container finished" podID="ea314c69-7524-43d6-9d4f-9fdb16510952" containerID="7a2cc30c7cfe8ca6acf3b5029da578678f6497bf01186f5ff5a4caf50683498c" exitCode=0 Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.173106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-qr6zk" event={"ID":"ea314c69-7524-43d6-9d4f-9fdb16510952","Type":"ContainerDied","Data":"7a2cc30c7cfe8ca6acf3b5029da578678f6497bf01186f5ff5a4caf50683498c"} Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.176024 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"0468589b613cbf08fc1bf01e9282b6c42aea8979704d41cdc9fd9d076d25db9d"} Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.708264 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.719868 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.740190 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849078 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-scripts\") pod \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849183 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pb5c\" (UniqueName: \"kubernetes.io/projected/6a1a6049-a924-4e9e-adad-e6ec84732eb9-kube-api-access-9pb5c\") pod \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849211 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-log-ovn\") pod \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849270 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rddff\" (UniqueName: \"kubernetes.io/projected/fa60a9bd-1933-44b8-ac17-be1f72c9da68-kube-api-access-rddff\") pod \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849297 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa60a9bd-1933-44b8-ac17-be1f72c9da68-operator-scripts\") pod \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\" (UID: \"fa60a9bd-1933-44b8-ac17-be1f72c9da68\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849312 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run\") pod \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849386 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drskx\" (UniqueName: \"kubernetes.io/projected/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-kube-api-access-drskx\") pod \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849409 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-additional-scripts\") pod \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849426 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run-ovn\") pod \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\" (UID: \"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.849467 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1a6049-a924-4e9e-adad-e6ec84732eb9-operator-scripts\") pod \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\" (UID: \"6a1a6049-a924-4e9e-adad-e6ec84732eb9\") " Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.850128 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa60a9bd-1933-44b8-ac17-be1f72c9da68-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fa60a9bd-1933-44b8-ac17-be1f72c9da68" (UID: "fa60a9bd-1933-44b8-ac17-be1f72c9da68"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.850352 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a1a6049-a924-4e9e-adad-e6ec84732eb9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a1a6049-a924-4e9e-adad-e6ec84732eb9" (UID: "6a1a6049-a924-4e9e-adad-e6ec84732eb9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.851077 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" (UID: "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.850967 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-scripts" (OuterVolumeSpecName: "scripts") pod "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" (UID: "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.850556 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run" (OuterVolumeSpecName: "var-run") pod "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" (UID: "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.851194 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" (UID: "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.851573 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" (UID: "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.853307 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a1a6049-a924-4e9e-adad-e6ec84732eb9-kube-api-access-9pb5c" (OuterVolumeSpecName: "kube-api-access-9pb5c") pod "6a1a6049-a924-4e9e-adad-e6ec84732eb9" (UID: "6a1a6049-a924-4e9e-adad-e6ec84732eb9"). InnerVolumeSpecName "kube-api-access-9pb5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.853622 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-kube-api-access-drskx" (OuterVolumeSpecName: "kube-api-access-drskx") pod "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" (UID: "1ce73342-b88b-4d1a-b6dc-8e7979f05e6d"). InnerVolumeSpecName "kube-api-access-drskx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.854020 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa60a9bd-1933-44b8-ac17-be1f72c9da68-kube-api-access-rddff" (OuterVolumeSpecName: "kube-api-access-rddff") pod "fa60a9bd-1933-44b8-ac17-be1f72c9da68" (UID: "fa60a9bd-1933-44b8-ac17-be1f72c9da68"). InnerVolumeSpecName "kube-api-access-rddff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950752 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drskx\" (UniqueName: \"kubernetes.io/projected/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-kube-api-access-drskx\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950804 4926 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950816 4926 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950827 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a1a6049-a924-4e9e-adad-e6ec84732eb9-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950836 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950845 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pb5c\" (UniqueName: \"kubernetes.io/projected/6a1a6049-a924-4e9e-adad-e6ec84732eb9-kube-api-access-9pb5c\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950853 4926 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950861 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rddff\" (UniqueName: \"kubernetes.io/projected/fa60a9bd-1933-44b8-ac17-be1f72c9da68-kube-api-access-rddff\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950869 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa60a9bd-1933-44b8-ac17-be1f72c9da68-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:33 crc kubenswrapper[4926]: I0312 18:22:33.950878 4926 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d-var-run\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.208365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-40f3-account-create-update-6f5kz" event={"ID":"6a1a6049-a924-4e9e-adad-e6ec84732eb9","Type":"ContainerDied","Data":"7f60838faf0603a5159f5f3ca03ddcba8992db0a0a7cb8b0a847b3f2ddc0a013"} Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.208675 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f60838faf0603a5159f5f3ca03ddcba8992db0a0a7cb8b0a847b3f2ddc0a013" Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.208389 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-40f3-account-create-update-6f5kz" Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.211859 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-n9jj2" event={"ID":"fa60a9bd-1933-44b8-ac17-be1f72c9da68","Type":"ContainerDied","Data":"2b292237feadafee95b31e06cfc840f72c515e88b8454c8337e1e74e28cd7576"} Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.211915 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b292237feadafee95b31e06cfc840f72c515e88b8454c8337e1e74e28cd7576" Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.211981 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-n9jj2" Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.224181 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-sfwpr-config-55drx" event={"ID":"1ce73342-b88b-4d1a-b6dc-8e7979f05e6d","Type":"ContainerDied","Data":"1b64556f5bf261422bd7a946c6906ac510a7650f9110c56570560ce6cb0ceff7"} Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.224229 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b64556f5bf261422bd7a946c6906ac510a7650f9110c56570560ce6cb0ceff7" Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.224291 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-sfwpr-config-55drx" Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.238275 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"f7925e6f8dd4e91bc33ea9de7de46827a465fec80fdc2c72f0e0c6ca7014c6b6"} Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.238691 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"98967e9369f3695e123659b5825995f560be830c73493e6998ebd0fb99612e49"} Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.811913 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-sfwpr-config-55drx"] Mar 12 18:22:34 crc kubenswrapper[4926]: I0312 18:22:34.823984 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-sfwpr-config-55drx"] Mar 12 18:22:36 crc kubenswrapper[4926]: I0312 18:22:36.503674 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" path="/var/lib/kubelet/pods/1ce73342-b88b-4d1a-b6dc-8e7979f05e6d/volumes" Mar 12 18:22:36 crc kubenswrapper[4926]: I0312 18:22:36.869217 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:36 crc kubenswrapper[4926]: I0312 18:22:36.913285 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:36 crc kubenswrapper[4926]: I0312 18:22:36.954303 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:36 crc kubenswrapper[4926]: I0312 18:22:36.983531 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.000720 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-kube-api-access-8nh4b\") pod \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.001129 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r6dd\" (UniqueName: \"kubernetes.io/projected/f7917312-5922-44df-a838-6b5452e8bb84-kube-api-access-7r6dd\") pod \"f7917312-5922-44df-a838-6b5452e8bb84\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.002388 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7917312-5922-44df-a838-6b5452e8bb84-operator-scripts\") pod \"f7917312-5922-44df-a838-6b5452e8bb84\" (UID: \"f7917312-5922-44df-a838-6b5452e8bb84\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.002455 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-operator-scripts\") pod \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\" (UID: \"9e5bf33c-5cb6-4b29-95f4-fcf47183be58\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.003737 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e5bf33c-5cb6-4b29-95f4-fcf47183be58" (UID: "9e5bf33c-5cb6-4b29-95f4-fcf47183be58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.004173 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7917312-5922-44df-a838-6b5452e8bb84-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f7917312-5922-44df-a838-6b5452e8bb84" (UID: "f7917312-5922-44df-a838-6b5452e8bb84"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.006747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7917312-5922-44df-a838-6b5452e8bb84-kube-api-access-7r6dd" (OuterVolumeSpecName: "kube-api-access-7r6dd") pod "f7917312-5922-44df-a838-6b5452e8bb84" (UID: "f7917312-5922-44df-a838-6b5452e8bb84"). InnerVolumeSpecName "kube-api-access-7r6dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.006803 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-kube-api-access-8nh4b" (OuterVolumeSpecName: "kube-api-access-8nh4b") pod "9e5bf33c-5cb6-4b29-95f4-fcf47183be58" (UID: "9e5bf33c-5cb6-4b29-95f4-fcf47183be58"). InnerVolumeSpecName "kube-api-access-8nh4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.104030 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7tgz\" (UniqueName: \"kubernetes.io/projected/05959443-d099-4653-9736-7745ba1ce331-kube-api-access-w7tgz\") pod \"05959443-d099-4653-9736-7745ba1ce331\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.104131 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05959443-d099-4653-9736-7745ba1ce331-operator-scripts\") pod \"05959443-d099-4653-9736-7745ba1ce331\" (UID: \"05959443-d099-4653-9736-7745ba1ce331\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.104294 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jl7q\" (UniqueName: \"kubernetes.io/projected/ea314c69-7524-43d6-9d4f-9fdb16510952-kube-api-access-7jl7q\") pod \"ea314c69-7524-43d6-9d4f-9fdb16510952\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.104921 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea314c69-7524-43d6-9d4f-9fdb16510952-operator-scripts\") pod \"ea314c69-7524-43d6-9d4f-9fdb16510952\" (UID: \"ea314c69-7524-43d6-9d4f-9fdb16510952\") " Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.104841 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05959443-d099-4653-9736-7745ba1ce331-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "05959443-d099-4653-9736-7745ba1ce331" (UID: "05959443-d099-4653-9736-7745ba1ce331"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.105795 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05959443-d099-4653-9736-7745ba1ce331-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.105955 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r6dd\" (UniqueName: \"kubernetes.io/projected/f7917312-5922-44df-a838-6b5452e8bb84-kube-api-access-7r6dd\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.106032 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f7917312-5922-44df-a838-6b5452e8bb84-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.106100 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.106176 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nh4b\" (UniqueName: \"kubernetes.io/projected/9e5bf33c-5cb6-4b29-95f4-fcf47183be58-kube-api-access-8nh4b\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.105798 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea314c69-7524-43d6-9d4f-9fdb16510952-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea314c69-7524-43d6-9d4f-9fdb16510952" (UID: "ea314c69-7524-43d6-9d4f-9fdb16510952"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.107419 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea314c69-7524-43d6-9d4f-9fdb16510952-kube-api-access-7jl7q" (OuterVolumeSpecName: "kube-api-access-7jl7q") pod "ea314c69-7524-43d6-9d4f-9fdb16510952" (UID: "ea314c69-7524-43d6-9d4f-9fdb16510952"). InnerVolumeSpecName "kube-api-access-7jl7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.108869 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05959443-d099-4653-9736-7745ba1ce331-kube-api-access-w7tgz" (OuterVolumeSpecName: "kube-api-access-w7tgz") pod "05959443-d099-4653-9736-7745ba1ce331" (UID: "05959443-d099-4653-9736-7745ba1ce331"). InnerVolumeSpecName "kube-api-access-w7tgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.207280 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jl7q\" (UniqueName: \"kubernetes.io/projected/ea314c69-7524-43d6-9d4f-9fdb16510952-kube-api-access-7jl7q\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.207555 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea314c69-7524-43d6-9d4f-9fdb16510952-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.207620 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7tgz\" (UniqueName: \"kubernetes.io/projected/05959443-d099-4653-9736-7745ba1ce331-kube-api-access-w7tgz\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.266500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"b030a2d014f400d50cb89a2eda000385e992f52900539bdd3c6a7b659a8269b2"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.266557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"0ad8612950d373ec053303346d238f478f191d4b05be61aed64cadc630109fad"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.268975 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-n28hs" event={"ID":"05959443-d099-4653-9736-7745ba1ce331","Type":"ContainerDied","Data":"def54e7f3bca68561150fcadb909db84a99aaf5cfe2458f9499bbf1014d7ccf8"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.269008 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="def54e7f3bca68561150fcadb909db84a99aaf5cfe2458f9499bbf1014d7ccf8" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.269012 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-n28hs" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.270927 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4vtkq" event={"ID":"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d","Type":"ContainerStarted","Data":"7fd785662fcf1b6a3363e17fd1fb61ae49e121593b0c0273db1785eefe4c3db8"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.274749 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-2c76-account-create-update-wvgzd" event={"ID":"f7917312-5922-44df-a838-6b5452e8bb84","Type":"ContainerDied","Data":"63fccee8b612554f7b094c76fecdac4ac35f19b95dc6b9a6784a3936c615d9a5"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.274910 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63fccee8b612554f7b094c76fecdac4ac35f19b95dc6b9a6784a3936c615d9a5" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.275006 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-2c76-account-create-update-wvgzd" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.287849 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bf50-account-create-update-ckzh6" event={"ID":"9e5bf33c-5cb6-4b29-95f4-fcf47183be58","Type":"ContainerDied","Data":"c56db79b42dcfa1fc516e7e174210c09707255662ee7602092c441dc8b95945d"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.287887 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c56db79b42dcfa1fc516e7e174210c09707255662ee7602092c441dc8b95945d" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.287940 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf50-account-create-update-ckzh6" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.294281 4926 generic.go:334] "Generic (PLEG): container finished" podID="6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" containerID="f8a8e8e5fcda402042df2295b1c2cbd321aa245d9b7be946af5559093b56242a" exitCode=0 Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.294402 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72msf" event={"ID":"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8","Type":"ContainerDied","Data":"f8a8e8e5fcda402042df2295b1c2cbd321aa245d9b7be946af5559093b56242a"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.296802 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-qr6zk" event={"ID":"ea314c69-7524-43d6-9d4f-9fdb16510952","Type":"ContainerDied","Data":"c5f3bf41c45721fdd76098163bfbbd5fb8a4d739371fec426b54c95a8d7546f5"} Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.296834 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5f3bf41c45721fdd76098163bfbbd5fb8a4d739371fec426b54c95a8d7546f5" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.300088 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-qr6zk" Mar 12 18:22:37 crc kubenswrapper[4926]: I0312 18:22:37.310572 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-4vtkq" podStartSLOduration=3.358687545 podStartE2EDuration="8.310552681s" podCreationTimestamp="2026-03-12 18:22:29 +0000 UTC" firstStartedPulling="2026-03-12 18:22:31.739421746 +0000 UTC m=+1192.108048089" lastFinishedPulling="2026-03-12 18:22:36.691286892 +0000 UTC m=+1197.059913225" observedRunningTime="2026-03-12 18:22:37.304243645 +0000 UTC m=+1197.672869978" watchObservedRunningTime="2026-03-12 18:22:37.310552681 +0000 UTC m=+1197.679179044" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.653968 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72msf" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.731862 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-combined-ca-bundle\") pod \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.731911 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-db-sync-config-data\") pod \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.731981 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-config-data\") pod \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.732080 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9kjm\" (UniqueName: \"kubernetes.io/projected/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-kube-api-access-q9kjm\") pod \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\" (UID: \"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8\") " Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.735743 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-kube-api-access-q9kjm" (OuterVolumeSpecName: "kube-api-access-q9kjm") pod "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" (UID: "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8"). InnerVolumeSpecName "kube-api-access-q9kjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.735984 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" (UID: "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.755675 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" (UID: "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.799273 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-config-data" (OuterVolumeSpecName: "config-data") pod "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" (UID: "6bad0b5a-a817-45ec-9ebb-7b30d7492ed8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.834142 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.834179 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9kjm\" (UniqueName: \"kubernetes.io/projected/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-kube-api-access-q9kjm\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.834191 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:38 crc kubenswrapper[4926]: I0312 18:22:38.834202 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.331812 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-72msf" event={"ID":"6bad0b5a-a817-45ec-9ebb-7b30d7492ed8","Type":"ContainerDied","Data":"0b8074caa3b0c10eb9a072ab0513f4ac17eb3bec44cc3edb55302dc941d344c0"} Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.332093 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b8074caa3b0c10eb9a072ab0513f4ac17eb3bec44cc3edb55302dc941d344c0" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.332162 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-72msf" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.345546 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"dcd686948bb98223945c78239c5e2eade4b334fdb486f13f8183a786b11bcb77"} Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.345806 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"2c4216509b2e96a37dad4fb826f0c82a6f461387fbac81c65033023a3b846844"} Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.345877 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"2814dd88eb80db2962578d8009f8f05d069c5caf4e38d32a30827d4b23f367dc"} Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.798662 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-s8pxk"] Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799231 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a1a6049-a924-4e9e-adad-e6ec84732eb9" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799243 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a1a6049-a924-4e9e-adad-e6ec84732eb9" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799260 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e5bf33c-5cb6-4b29-95f4-fcf47183be58" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799266 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e5bf33c-5cb6-4b29-95f4-fcf47183be58" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799277 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea314c69-7524-43d6-9d4f-9fdb16510952" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799284 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea314c69-7524-43d6-9d4f-9fdb16510952" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799294 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="349ca4f5-349b-45ab-98a4-844fa00599d0" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799300 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="349ca4f5-349b-45ab-98a4-844fa00599d0" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799308 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7917312-5922-44df-a838-6b5452e8bb84" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799315 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7917312-5922-44df-a838-6b5452e8bb84" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799326 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05959443-d099-4653-9736-7745ba1ce331" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799331 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="05959443-d099-4653-9736-7745ba1ce331" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799342 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" containerName="ovn-config" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799347 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" containerName="ovn-config" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799362 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa60a9bd-1933-44b8-ac17-be1f72c9da68" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799368 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa60a9bd-1933-44b8-ac17-be1f72c9da68" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: E0312 18:22:39.799377 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" containerName="glance-db-sync" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.799383 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" containerName="glance-db-sync" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807497 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="349ca4f5-349b-45ab-98a4-844fa00599d0" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807542 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e5bf33c-5cb6-4b29-95f4-fcf47183be58" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807558 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea314c69-7524-43d6-9d4f-9fdb16510952" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807573 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7917312-5922-44df-a838-6b5452e8bb84" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807587 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" containerName="glance-db-sync" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807596 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a1a6049-a924-4e9e-adad-e6ec84732eb9" containerName="mariadb-account-create-update" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807616 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="05959443-d099-4653-9736-7745ba1ce331" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807632 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa60a9bd-1933-44b8-ac17-be1f72c9da68" containerName="mariadb-database-create" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.807641 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ce73342-b88b-4d1a-b6dc-8e7979f05e6d" containerName="ovn-config" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.808926 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.819009 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-s8pxk"] Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.952942 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-config\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.953047 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.953145 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.953241 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xrz5\" (UniqueName: \"kubernetes.io/projected/4f5dc206-65d8-4b00-a55d-b505d640a279-kube-api-access-9xrz5\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:39 crc kubenswrapper[4926]: I0312 18:22:39.953318 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.058240 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.058304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-config\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.058326 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.058395 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.058469 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xrz5\" (UniqueName: \"kubernetes.io/projected/4f5dc206-65d8-4b00-a55d-b505d640a279-kube-api-access-9xrz5\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.059909 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.060412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-config\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.060918 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.061465 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.076372 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xrz5\" (UniqueName: \"kubernetes.io/projected/4f5dc206-65d8-4b00-a55d-b505d640a279-kube-api-access-9xrz5\") pod \"dnsmasq-dns-5b946c75cc-s8pxk\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.127848 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.364120 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"3505cb804eb89707510f827ea55bf8c0b5498a603a4dc33a295007ca469842dc"} Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.364376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"1661ebf70a682193744c016efe51c609283a83267b967202bbe346b08fdc327c"} Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.364386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"35156daa5963895ee8abf10c0b8630bfb036d1245a6d593e55409209aabbf565"} Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.364396 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"57853681-32de-4475-9c7d-3f9708fe7d91","Type":"ContainerStarted","Data":"46bc1bb64318be0810542d3dd7cfbc08c1c22a5f45619d5edb2d27d87b4afabe"} Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.584876 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=38.789438171 podStartE2EDuration="48.584857654s" podCreationTimestamp="2026-03-12 18:21:52 +0000 UTC" firstStartedPulling="2026-03-12 18:22:28.50747937 +0000 UTC m=+1188.876105703" lastFinishedPulling="2026-03-12 18:22:38.302898843 +0000 UTC m=+1198.671525186" observedRunningTime="2026-03-12 18:22:40.407765026 +0000 UTC m=+1200.776391359" watchObservedRunningTime="2026-03-12 18:22:40.584857654 +0000 UTC m=+1200.953483987" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.587095 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-s8pxk"] Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.673373 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-s8pxk"] Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.713336 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tvrln"] Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.714994 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.717203 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.733967 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tvrln"] Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.766955 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.766998 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4dg8\" (UniqueName: \"kubernetes.io/projected/69fa4317-bafb-462d-b6aa-ea07437277f1-kube-api-access-p4dg8\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.767066 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.767113 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.767155 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.767180 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-config\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.869181 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.869262 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.869315 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-config\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.869451 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.869490 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4dg8\" (UniqueName: \"kubernetes.io/projected/69fa4317-bafb-462d-b6aa-ea07437277f1-kube-api-access-p4dg8\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.869530 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.870886 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.870897 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.870951 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.870890 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.871028 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-config\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:40 crc kubenswrapper[4926]: I0312 18:22:40.889038 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4dg8\" (UniqueName: \"kubernetes.io/projected/69fa4317-bafb-462d-b6aa-ea07437277f1-kube-api-access-p4dg8\") pod \"dnsmasq-dns-74f6bcbc87-tvrln\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:41 crc kubenswrapper[4926]: I0312 18:22:41.072139 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:41 crc kubenswrapper[4926]: I0312 18:22:41.372712 4926 generic.go:334] "Generic (PLEG): container finished" podID="4f5dc206-65d8-4b00-a55d-b505d640a279" containerID="be98fd8613351be05b0464caa0d08812f4423f1da0f0227e98091b428e4425a5" exitCode=0 Mar 12 18:22:41 crc kubenswrapper[4926]: I0312 18:22:41.372793 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" event={"ID":"4f5dc206-65d8-4b00-a55d-b505d640a279","Type":"ContainerDied","Data":"be98fd8613351be05b0464caa0d08812f4423f1da0f0227e98091b428e4425a5"} Mar 12 18:22:41 crc kubenswrapper[4926]: I0312 18:22:41.373025 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" event={"ID":"4f5dc206-65d8-4b00-a55d-b505d640a279","Type":"ContainerStarted","Data":"62587664fe08f5a1f36b2ec950757fbec287bd8897e920da359c13c46fb57219"} Mar 12 18:22:41 crc kubenswrapper[4926]: I0312 18:22:41.374597 4926 generic.go:334] "Generic (PLEG): container finished" podID="d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" containerID="7fd785662fcf1b6a3363e17fd1fb61ae49e121593b0c0273db1785eefe4c3db8" exitCode=0 Mar 12 18:22:41 crc kubenswrapper[4926]: I0312 18:22:41.374666 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4vtkq" event={"ID":"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d","Type":"ContainerDied","Data":"7fd785662fcf1b6a3363e17fd1fb61ae49e121593b0c0273db1785eefe4c3db8"} Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.051702 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.119393 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tvrln"] Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.124354 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xrz5\" (UniqueName: \"kubernetes.io/projected/4f5dc206-65d8-4b00-a55d-b505d640a279-kube-api-access-9xrz5\") pod \"4f5dc206-65d8-4b00-a55d-b505d640a279\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.124478 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-sb\") pod \"4f5dc206-65d8-4b00-a55d-b505d640a279\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.124523 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-nb\") pod \"4f5dc206-65d8-4b00-a55d-b505d640a279\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.124542 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-dns-svc\") pod \"4f5dc206-65d8-4b00-a55d-b505d640a279\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.124575 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-config\") pod \"4f5dc206-65d8-4b00-a55d-b505d640a279\" (UID: \"4f5dc206-65d8-4b00-a55d-b505d640a279\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.132043 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f5dc206-65d8-4b00-a55d-b505d640a279-kube-api-access-9xrz5" (OuterVolumeSpecName: "kube-api-access-9xrz5") pod "4f5dc206-65d8-4b00-a55d-b505d640a279" (UID: "4f5dc206-65d8-4b00-a55d-b505d640a279"). InnerVolumeSpecName "kube-api-access-9xrz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.146607 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f5dc206-65d8-4b00-a55d-b505d640a279" (UID: "4f5dc206-65d8-4b00-a55d-b505d640a279"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.148243 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-config" (OuterVolumeSpecName: "config") pod "4f5dc206-65d8-4b00-a55d-b505d640a279" (UID: "4f5dc206-65d8-4b00-a55d-b505d640a279"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.150570 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f5dc206-65d8-4b00-a55d-b505d640a279" (UID: "4f5dc206-65d8-4b00-a55d-b505d640a279"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.155265 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f5dc206-65d8-4b00-a55d-b505d640a279" (UID: "4f5dc206-65d8-4b00-a55d-b505d640a279"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.232486 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.232525 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.232540 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.232553 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f5dc206-65d8-4b00-a55d-b505d640a279-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.232567 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xrz5\" (UniqueName: \"kubernetes.io/projected/4f5dc206-65d8-4b00-a55d-b505d640a279-kube-api-access-9xrz5\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.387666 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.387673 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-s8pxk" event={"ID":"4f5dc206-65d8-4b00-a55d-b505d640a279","Type":"ContainerDied","Data":"62587664fe08f5a1f36b2ec950757fbec287bd8897e920da359c13c46fb57219"} Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.387826 4926 scope.go:117] "RemoveContainer" containerID="be98fd8613351be05b0464caa0d08812f4423f1da0f0227e98091b428e4425a5" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.391585 4926 generic.go:334] "Generic (PLEG): container finished" podID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerID="fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b" exitCode=0 Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.392619 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" event={"ID":"69fa4317-bafb-462d-b6aa-ea07437277f1","Type":"ContainerDied","Data":"fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b"} Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.392656 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" event={"ID":"69fa4317-bafb-462d-b6aa-ea07437277f1","Type":"ContainerStarted","Data":"a0cec62e2c7e846902634b014c1304ee38e6c7bc3f234649898d90b9aefa0ac7"} Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.492373 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-s8pxk"] Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.504130 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-s8pxk"] Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.725289 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.843993 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-config-data\") pod \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.844165 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8k85\" (UniqueName: \"kubernetes.io/projected/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-kube-api-access-w8k85\") pod \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.844290 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-combined-ca-bundle\") pod \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\" (UID: \"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d\") " Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.847542 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-kube-api-access-w8k85" (OuterVolumeSpecName: "kube-api-access-w8k85") pod "d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" (UID: "d7b6b9d3-0c3c-4dcc-b417-49d53269c39d"). InnerVolumeSpecName "kube-api-access-w8k85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.868194 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" (UID: "d7b6b9d3-0c3c-4dcc-b417-49d53269c39d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.885765 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-config-data" (OuterVolumeSpecName: "config-data") pod "d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" (UID: "d7b6b9d3-0c3c-4dcc-b417-49d53269c39d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.946599 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.946639 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8k85\" (UniqueName: \"kubernetes.io/projected/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-kube-api-access-w8k85\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:42 crc kubenswrapper[4926]: I0312 18:22:42.946657 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.402036 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" event={"ID":"69fa4317-bafb-462d-b6aa-ea07437277f1","Type":"ContainerStarted","Data":"9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee"} Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.403267 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.412157 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4vtkq" event={"ID":"d7b6b9d3-0c3c-4dcc-b417-49d53269c39d","Type":"ContainerDied","Data":"7c7e1f1cb85b710df8d9fc682c9ab5a089b0604e2b56bc8fd5036a48bcce3088"} Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.412197 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c7e1f1cb85b710df8d9fc682c9ab5a089b0604e2b56bc8fd5036a48bcce3088" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.412223 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4vtkq" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.435801 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" podStartSLOduration=3.4357776700000002 podStartE2EDuration="3.43577767s" podCreationTimestamp="2026-03-12 18:22:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:43.430453844 +0000 UTC m=+1203.799080177" watchObservedRunningTime="2026-03-12 18:22:43.43577767 +0000 UTC m=+1203.804404003" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.682565 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tvrln"] Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.717873 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hrv24"] Mar 12 18:22:43 crc kubenswrapper[4926]: E0312 18:22:43.718486 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" containerName="keystone-db-sync" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.718569 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" containerName="keystone-db-sync" Mar 12 18:22:43 crc kubenswrapper[4926]: E0312 18:22:43.718642 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f5dc206-65d8-4b00-a55d-b505d640a279" containerName="init" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.718716 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f5dc206-65d8-4b00-a55d-b505d640a279" containerName="init" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.718968 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" containerName="keystone-db-sync" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.719050 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f5dc206-65d8-4b00-a55d-b505d640a279" containerName="init" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.725772 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.732614 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.732951 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.733173 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vrsc7" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.733246 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.733466 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.807224 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6j58l"] Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.808474 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.828533 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hrv24"] Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.841613 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6j58l"] Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.877315 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-config-data\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.877399 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-combined-ca-bundle\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.877429 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-fernet-keys\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.877461 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-credential-keys\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.877486 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpnk7\" (UniqueName: \"kubernetes.io/projected/fdce3e2c-d108-4904-b4ab-8d95d4838bad-kube-api-access-rpnk7\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.877628 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-scripts\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980604 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-fernet-keys\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980655 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-credential-keys\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980678 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpnk7\" (UniqueName: \"kubernetes.io/projected/fdce3e2c-d108-4904-b4ab-8d95d4838bad-kube-api-access-rpnk7\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980720 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-config\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980792 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-scripts\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980816 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-config-data\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980908 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980936 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980970 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-combined-ca-bundle\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.980992 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:43 crc kubenswrapper[4926]: I0312 18:22:43.981016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sftv\" (UniqueName: \"kubernetes.io/projected/c72142af-cd95-4044-92d7-ce173109afa7-kube-api-access-4sftv\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.007132 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-credential-keys\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.007845 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-fernet-keys\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.008329 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-config-data\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.021993 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7786dc65fc-n6hp6"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.023233 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.031164 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-combined-ca-bundle\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.031387 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-scripts\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.032620 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.032865 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.033680 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.048706 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-4vqq4" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.053574 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7786dc65fc-n6hp6"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.093245 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-config\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.093332 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.093390 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.093411 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.093447 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.093465 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sftv\" (UniqueName: \"kubernetes.io/projected/c72142af-cd95-4044-92d7-ce173109afa7-kube-api-access-4sftv\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.094584 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-config\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.095100 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.095597 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.096110 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.103157 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.103934 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.118973 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpnk7\" (UniqueName: \"kubernetes.io/projected/fdce3e2c-d108-4904-b4ab-8d95d4838bad-kube-api-access-rpnk7\") pod \"keystone-bootstrap-hrv24\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.133885 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.153596 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.153781 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.172210 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.192146 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sftv\" (UniqueName: \"kubernetes.io/projected/c72142af-cd95-4044-92d7-ce173109afa7-kube-api-access-4sftv\") pod \"dnsmasq-dns-847c4cc679-6j58l\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.195369 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2a8323-527c-4e19-ab55-3c291c4d538f-logs\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.195422 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c2a8323-527c-4e19-ab55-3c291c4d538f-horizon-secret-key\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.195463 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-scripts\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.195512 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-config-data\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.195601 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flm2j\" (UniqueName: \"kubernetes.io/projected/8c2a8323-527c-4e19-ab55-3c291c4d538f-kube-api-access-flm2j\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.235620 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-vc5cr"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.236988 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.245005 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.253999 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.254251 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qpbhw" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.254751 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-sgnbr"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.255761 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.271106 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hz28b" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.271359 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.296486 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vc5cr"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297351 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2a8323-527c-4e19-ab55-3c291c4d538f-logs\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297390 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-scripts\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297407 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-config-data\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297449 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c2a8323-527c-4e19-ab55-3c291c4d538f-horizon-secret-key\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297471 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-scripts\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297496 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smrs2\" (UniqueName: \"kubernetes.io/projected/5fdfffa4-937c-4167-8545-d34f2007fbc9-kube-api-access-smrs2\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297526 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-config-data\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297542 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297562 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297583 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flm2j\" (UniqueName: \"kubernetes.io/projected/8c2a8323-527c-4e19-ab55-3c291c4d538f-kube-api-access-flm2j\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297614 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-run-httpd\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297662 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-log-httpd\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.297922 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2a8323-527c-4e19-ab55-3c291c4d538f-logs\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.298514 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-scripts\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.299473 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-config-data\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.327526 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-txt96"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.328616 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.340399 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c2a8323-527c-4e19-ab55-3c291c4d538f-horizon-secret-key\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.340616 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.349170 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.352299 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flm2j\" (UniqueName: \"kubernetes.io/projected/8c2a8323-527c-4e19-ab55-3c291c4d538f-kube-api-access-flm2j\") pod \"horizon-7786dc65fc-n6hp6\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.352414 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sgnbr"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.360726 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.367955 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9lh8l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.384708 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-txt96"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402697 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-combined-ca-bundle\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402765 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df6kk\" (UniqueName: \"kubernetes.io/projected/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-kube-api-access-df6kk\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402787 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-config-data\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402805 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgd7q\" (UniqueName: \"kubernetes.io/projected/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-kube-api-access-tgd7q\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402826 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-db-sync-config-data\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402846 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-config\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-combined-ca-bundle\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402883 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-scripts\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402902 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-scripts\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-config-data\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402951 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mklss\" (UniqueName: \"kubernetes.io/projected/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-kube-api-access-mklss\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.402980 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smrs2\" (UniqueName: \"kubernetes.io/projected/5fdfffa4-937c-4167-8545-d34f2007fbc9-kube-api-access-smrs2\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.403016 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.403040 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.403062 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-combined-ca-bundle\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.403091 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-logs\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.403112 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-run-httpd\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.403130 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-log-httpd\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.403552 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-log-httpd\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.413230 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-run-httpd\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.418017 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-lgvzs"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.419082 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.426970 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.427245 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.427511 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-m6xv4" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.429353 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.432024 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.433280 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-scripts\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.441920 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-config-data\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.449803 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.473400 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.481411 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-84786596b5-28lbk"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.482732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.508830 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgqfm\" (UniqueName: \"kubernetes.io/projected/dac4b5d6-fb31-4955-8679-db9d3ff63c10-kube-api-access-fgqfm\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.508869 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mklss\" (UniqueName: \"kubernetes.io/projected/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-kube-api-access-mklss\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.508894 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-config-data\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.508917 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-combined-ca-bundle\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.508956 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-combined-ca-bundle\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.508976 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-scripts\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509000 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-logs\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509037 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-combined-ca-bundle\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509055 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dac4b5d6-fb31-4955-8679-db9d3ff63c10-etc-machine-id\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509078 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df6kk\" (UniqueName: \"kubernetes.io/projected/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-kube-api-access-df6kk\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509093 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-config-data\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509111 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-db-sync-config-data\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509128 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgd7q\" (UniqueName: \"kubernetes.io/projected/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-kube-api-access-tgd7q\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-db-sync-config-data\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509169 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-config\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509185 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-combined-ca-bundle\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509202 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-scripts\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.509206 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smrs2\" (UniqueName: \"kubernetes.io/projected/5fdfffa4-937c-4167-8545-d34f2007fbc9-kube-api-access-smrs2\") pod \"ceilometer-0\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.512239 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-logs\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.520081 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-db-sync-config-data\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.525341 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f5dc206-65d8-4b00-a55d-b505d640a279" path="/var/lib/kubelet/pods/4f5dc206-65d8-4b00-a55d-b505d640a279/volumes" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.533672 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-scripts\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.534865 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-combined-ca-bundle\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.537362 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgd7q\" (UniqueName: \"kubernetes.io/projected/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-kube-api-access-tgd7q\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.538025 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-config-data\") pod \"placement-db-sync-txt96\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.538674 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-config\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.541079 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-combined-ca-bundle\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.546050 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-combined-ca-bundle\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.551039 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6j58l"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.556800 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mklss\" (UniqueName: \"kubernetes.io/projected/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-kube-api-access-mklss\") pod \"barbican-db-sync-sgnbr\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.572950 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df6kk\" (UniqueName: \"kubernetes.io/projected/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-kube-api-access-df6kk\") pod \"neutron-db-sync-vc5cr\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.601156 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-lgvzs"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.610629 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84786596b5-28lbk"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611427 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7bxq\" (UniqueName: \"kubernetes.io/projected/1ca9ea6f-908c-48db-9313-c3ff4809a993-kube-api-access-p7bxq\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611486 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ca9ea6f-908c-48db-9313-c3ff4809a993-logs\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611521 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgqfm\" (UniqueName: \"kubernetes.io/projected/dac4b5d6-fb31-4955-8679-db9d3ff63c10-kube-api-access-fgqfm\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-scripts\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611615 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-config-data\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611641 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-combined-ca-bundle\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611667 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-config-data\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611733 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-scripts\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611823 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ca9ea6f-908c-48db-9313-c3ff4809a993-horizon-secret-key\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611889 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dac4b5d6-fb31-4955-8679-db9d3ff63c10-etc-machine-id\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.611940 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-db-sync-config-data\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.613865 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dac4b5d6-fb31-4955-8679-db9d3ff63c10-etc-machine-id\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.615486 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-db-sync-config-data\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.619080 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-combined-ca-bundle\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.622392 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-config-data\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.622947 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-scripts\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.629955 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7pqqh"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.632341 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.672400 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgqfm\" (UniqueName: \"kubernetes.io/projected/dac4b5d6-fb31-4955-8679-db9d3ff63c10-kube-api-access-fgqfm\") pod \"cinder-db-sync-lgvzs\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.704968 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7pqqh"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.717048 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-txt96" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.717248 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.718748 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ca9ea6f-908c-48db-9313-c3ff4809a993-horizon-secret-key\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.718801 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.718849 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7bxq\" (UniqueName: \"kubernetes.io/projected/1ca9ea6f-908c-48db-9313-c3ff4809a993-kube-api-access-p7bxq\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.718882 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.718921 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ca9ea6f-908c-48db-9313-c3ff4809a993-logs\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.718945 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-scripts\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.718989 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.719013 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.719028 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.719051 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzlwc\" (UniqueName: \"kubernetes.io/projected/af5b15ef-fa73-4ef4-9235-41da81503d2c-kube-api-access-gzlwc\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.719071 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-config-data\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.719806 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ca9ea6f-908c-48db-9313-c3ff4809a993-logs\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.720639 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-config-data\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.720884 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.724277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-scripts\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.728201 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ca9ea6f-908c-48db-9313-c3ff4809a993-horizon-secret-key\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.731937 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.735318 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.737810 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2zfbp" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.738084 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.759044 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.760229 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.762375 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7bxq\" (UniqueName: \"kubernetes.io/projected/1ca9ea6f-908c-48db-9313-c3ff4809a993-kube-api-access-p7bxq\") pod \"horizon-84786596b5-28lbk\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.773774 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.776356 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.805021 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.816779 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.818461 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.821108 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.821176 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.821229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.821247 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.821476 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.821586 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzlwc\" (UniqueName: \"kubernetes.io/projected/af5b15ef-fa73-4ef4-9235-41da81503d2c-kube-api-access-gzlwc\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.822616 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.823839 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.825338 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.827004 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.827180 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.828186 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.830210 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.837853 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.841136 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.848049 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzlwc\" (UniqueName: \"kubernetes.io/projected/af5b15ef-fa73-4ef4-9235-41da81503d2c-kube-api-access-gzlwc\") pod \"dnsmasq-dns-785d8bcb8c-7pqqh\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924559 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924608 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924649 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924669 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-logs\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924711 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924730 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn68d\" (UniqueName: \"kubernetes.io/projected/7a40ca3d-194d-4baa-8417-3614c8aeef08-kube-api-access-xn68d\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924764 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924783 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924814 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924845 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924886 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-config-data\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924915 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-logs\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924947 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2njd\" (UniqueName: \"kubernetes.io/projected/2900c0f0-0ed1-493a-8bad-ed203682251d-kube-api-access-z2njd\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.924963 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-scripts\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.925636 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:44 crc kubenswrapper[4926]: I0312 18:22:44.925712 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.029775 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030252 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-logs\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030280 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030294 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn68d\" (UniqueName: \"kubernetes.io/projected/7a40ca3d-194d-4baa-8417-3614c8aeef08-kube-api-access-xn68d\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030315 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030333 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030381 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030428 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030505 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-config-data\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030537 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-logs\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030579 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2njd\" (UniqueName: \"kubernetes.io/projected/2900c0f0-0ed1-493a-8bad-ed203682251d-kube-api-access-z2njd\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030602 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-scripts\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030637 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030664 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030690 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030714 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.030932 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-logs\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.031184 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.031521 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.031788 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-logs\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.032039 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.046237 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-config-data\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.047063 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.047746 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.048031 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-scripts\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.048094 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.052413 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.053430 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn68d\" (UniqueName: \"kubernetes.io/projected/7a40ca3d-194d-4baa-8417-3614c8aeef08-kube-api-access-xn68d\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.055688 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.057418 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.064798 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.069571 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.071298 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2njd\" (UniqueName: \"kubernetes.io/projected/2900c0f0-0ed1-493a-8bad-ed203682251d-kube-api-access-z2njd\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.081077 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.083430 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.086052 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6j58l"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.154242 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.227470 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hrv24"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.259226 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7786dc65fc-n6hp6"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.376133 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.458826 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hrv24" event={"ID":"fdce3e2c-d108-4904-b4ab-8d95d4838bad","Type":"ContainerStarted","Data":"334e13044642c4b0a613f2bfccdc8e4e0067f4f734093e62a567052168539e3c"} Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.459150 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hrv24" event={"ID":"fdce3e2c-d108-4904-b4ab-8d95d4838bad","Type":"ContainerStarted","Data":"cbd223739736371712c30899d1265bcdf31be415ef718a7391e85bb96801c5f5"} Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.460995 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7786dc65fc-n6hp6" event={"ID":"8c2a8323-527c-4e19-ab55-3c291c4d538f","Type":"ContainerStarted","Data":"1a2407e300165d17a9ee94f4b0ab357bc8d93ae3d9e96138519ce22a7544a84a"} Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.462967 4926 generic.go:334] "Generic (PLEG): container finished" podID="c72142af-cd95-4044-92d7-ce173109afa7" containerID="05f95959f0e95ab682be03e1080674ba4ddff508d0c0201c65a7316a98a66c30" exitCode=0 Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.463172 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" podUID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerName="dnsmasq-dns" containerID="cri-o://9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee" gracePeriod=10 Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.463611 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6j58l" event={"ID":"c72142af-cd95-4044-92d7-ce173109afa7","Type":"ContainerDied","Data":"05f95959f0e95ab682be03e1080674ba4ddff508d0c0201c65a7316a98a66c30"} Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.463651 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6j58l" event={"ID":"c72142af-cd95-4044-92d7-ce173109afa7","Type":"ContainerStarted","Data":"cac6ba4747f07ad7b159a2c64c5d5c743305d616dd8e88a4a78b6bbef0d66fcd"} Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.482250 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hrv24" podStartSLOduration=2.482226876 podStartE2EDuration="2.482226876s" podCreationTimestamp="2026-03-12 18:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:45.479236353 +0000 UTC m=+1205.847862686" watchObservedRunningTime="2026-03-12 18:22:45.482226876 +0000 UTC m=+1205.850853219" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.715651 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vc5cr"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.743073 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sgnbr"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.752566 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-84786596b5-28lbk"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.764246 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-lgvzs"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.774240 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-txt96"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.812776 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.821718 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7pqqh"] Mar 12 18:22:45 crc kubenswrapper[4926]: W0312 18:22:45.823974 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf5b15ef_fa73_4ef4_9235_41da81503d2c.slice/crio-c327e0d0fbf885a0bb6f202d846326d2682c20d15307d1ed071e6aa3068c47c0 WatchSource:0}: Error finding container c327e0d0fbf885a0bb6f202d846326d2682c20d15307d1ed071e6aa3068c47c0: Status 404 returned error can't find the container with id c327e0d0fbf885a0bb6f202d846326d2682c20d15307d1ed071e6aa3068c47c0 Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.910297 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.966580 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-svc\") pod \"c72142af-cd95-4044-92d7-ce173109afa7\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.966738 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-sb\") pod \"c72142af-cd95-4044-92d7-ce173109afa7\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.966830 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sftv\" (UniqueName: \"kubernetes.io/projected/c72142af-cd95-4044-92d7-ce173109afa7-kube-api-access-4sftv\") pod \"c72142af-cd95-4044-92d7-ce173109afa7\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.966982 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-config\") pod \"c72142af-cd95-4044-92d7-ce173109afa7\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.968799 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-swift-storage-0\") pod \"c72142af-cd95-4044-92d7-ce173109afa7\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.968854 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-nb\") pod \"c72142af-cd95-4044-92d7-ce173109afa7\" (UID: \"c72142af-cd95-4044-92d7-ce173109afa7\") " Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.993811 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c72142af-cd95-4044-92d7-ce173109afa7-kube-api-access-4sftv" (OuterVolumeSpecName: "kube-api-access-4sftv") pod "c72142af-cd95-4044-92d7-ce173109afa7" (UID: "c72142af-cd95-4044-92d7-ce173109afa7"). InnerVolumeSpecName "kube-api-access-4sftv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:45 crc kubenswrapper[4926]: I0312 18:22:45.996299 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c72142af-cd95-4044-92d7-ce173109afa7" (UID: "c72142af-cd95-4044-92d7-ce173109afa7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.011191 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c72142af-cd95-4044-92d7-ce173109afa7" (UID: "c72142af-cd95-4044-92d7-ce173109afa7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.027244 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-config" (OuterVolumeSpecName: "config") pod "c72142af-cd95-4044-92d7-ce173109afa7" (UID: "c72142af-cd95-4044-92d7-ce173109afa7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.040759 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c72142af-cd95-4044-92d7-ce173109afa7" (UID: "c72142af-cd95-4044-92d7-ce173109afa7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.060521 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c72142af-cd95-4044-92d7-ce173109afa7" (UID: "c72142af-cd95-4044-92d7-ce173109afa7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.077730 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.077770 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.077784 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sftv\" (UniqueName: \"kubernetes.io/projected/c72142af-cd95-4044-92d7-ce173109afa7-kube-api-access-4sftv\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.077794 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.077802 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.077810 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c72142af-cd95-4044-92d7-ce173109afa7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.167468 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.257081 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7786dc65fc-n6hp6"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.283339 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-749fd9494c-gtzjf"] Mar 12 18:22:46 crc kubenswrapper[4926]: E0312 18:22:46.283679 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c72142af-cd95-4044-92d7-ce173109afa7" containerName="init" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.283696 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c72142af-cd95-4044-92d7-ce173109afa7" containerName="init" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.283847 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c72142af-cd95-4044-92d7-ce173109afa7" containerName="init" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.284698 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.294752 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.306761 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-749fd9494c-gtzjf"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.326805 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.365519 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.387108 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-nb\") pod \"69fa4317-bafb-462d-b6aa-ea07437277f1\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389089 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-svc\") pod \"69fa4317-bafb-462d-b6aa-ea07437277f1\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389135 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-swift-storage-0\") pod \"69fa4317-bafb-462d-b6aa-ea07437277f1\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389172 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-sb\") pod \"69fa4317-bafb-462d-b6aa-ea07437277f1\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389248 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-config\") pod \"69fa4317-bafb-462d-b6aa-ea07437277f1\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389287 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4dg8\" (UniqueName: \"kubernetes.io/projected/69fa4317-bafb-462d-b6aa-ea07437277f1-kube-api-access-p4dg8\") pod \"69fa4317-bafb-462d-b6aa-ea07437277f1\" (UID: \"69fa4317-bafb-462d-b6aa-ea07437277f1\") " Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389710 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-config-data\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389776 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-scripts\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.389856 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04df6e00-4552-471c-9ae0-f45362e7e2b4-logs\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.390008 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04df6e00-4552-471c-9ae0-f45362e7e2b4-horizon-secret-key\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.390036 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g599\" (UniqueName: \"kubernetes.io/projected/04df6e00-4552-471c-9ae0-f45362e7e2b4-kube-api-access-8g599\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.399634 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69fa4317-bafb-462d-b6aa-ea07437277f1-kube-api-access-p4dg8" (OuterVolumeSpecName: "kube-api-access-p4dg8") pod "69fa4317-bafb-462d-b6aa-ea07437277f1" (UID: "69fa4317-bafb-462d-b6aa-ea07437277f1"). InnerVolumeSpecName "kube-api-access-p4dg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.437309 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.449472 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.493212 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04df6e00-4552-471c-9ae0-f45362e7e2b4-horizon-secret-key\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.493656 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g599\" (UniqueName: \"kubernetes.io/projected/04df6e00-4552-471c-9ae0-f45362e7e2b4-kube-api-access-8g599\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.493800 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-config-data\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.493859 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-scripts\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.493940 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04df6e00-4552-471c-9ae0-f45362e7e2b4-logs\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.494072 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4dg8\" (UniqueName: \"kubernetes.io/projected/69fa4317-bafb-462d-b6aa-ea07437277f1-kube-api-access-p4dg8\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.494676 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04df6e00-4552-471c-9ae0-f45362e7e2b4-logs\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.497198 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-config-data\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.497981 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-scripts\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.513370 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04df6e00-4552-471c-9ae0-f45362e7e2b4-horizon-secret-key\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.519375 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g599\" (UniqueName: \"kubernetes.io/projected/04df6e00-4552-471c-9ae0-f45362e7e2b4-kube-api-access-8g599\") pod \"horizon-749fd9494c-gtzjf\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.536092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lgvzs" event={"ID":"dac4b5d6-fb31-4955-8679-db9d3ff63c10","Type":"ContainerStarted","Data":"1b929dd828244692447588d25c4d9ac5aab4e983cee4fcf89d743277f69d1b5d"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.536141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84786596b5-28lbk" event={"ID":"1ca9ea6f-908c-48db-9313-c3ff4809a993","Type":"ContainerStarted","Data":"87827b8a20d43f6415fa4987b8da56ced353164794c7d29438fb6ffa36394a39"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.536156 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2900c0f0-0ed1-493a-8bad-ed203682251d","Type":"ContainerStarted","Data":"7a86c1cac56bc2792ca8a01485262ffdadcde243619bf97536386064fe81ed59"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.536170 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" event={"ID":"af5b15ef-fa73-4ef4-9235-41da81503d2c","Type":"ContainerStarted","Data":"c327e0d0fbf885a0bb6f202d846326d2682c20d15307d1ed071e6aa3068c47c0"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.536182 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sgnbr" event={"ID":"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6","Type":"ContainerStarted","Data":"89660f246e3c30bdbd43247de70865e18df164fd93b36ad1830555d68b2f5032"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.536194 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vc5cr" event={"ID":"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b","Type":"ContainerStarted","Data":"d40939b357197123b7378525a59c3f91e7240cb30140873883a85a0672ad25de"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.536206 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vc5cr" event={"ID":"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b","Type":"ContainerStarted","Data":"85be7b1fc6b389d0d4f2769e75e7678715b75c3e095cab19a3e0aad9bb5656b7"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.539547 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerStarted","Data":"fde586bc97eba127841e218475be2e383628bdf6feeb7595fc6e0b0daaf96628"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.546874 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-txt96" event={"ID":"7a0f2830-bf50-4195-9dad-d4d2c9529ee9","Type":"ContainerStarted","Data":"ce15d4b03307ac2a9a7f3c1d05eba2559878bf699c0483537cc9d838c8197125"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.552807 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-6j58l" event={"ID":"c72142af-cd95-4044-92d7-ce173109afa7","Type":"ContainerDied","Data":"cac6ba4747f07ad7b159a2c64c5d5c743305d616dd8e88a4a78b6bbef0d66fcd"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.552863 4926 scope.go:117] "RemoveContainer" containerID="05f95959f0e95ab682be03e1080674ba4ddff508d0c0201c65a7316a98a66c30" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.552983 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-6j58l" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.562448 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a40ca3d-194d-4baa-8417-3614c8aeef08","Type":"ContainerStarted","Data":"d093554ce00472559ced5d4b721fe022998a063b108c7d7113b402e4a98bca54"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.568257 4926 generic.go:334] "Generic (PLEG): container finished" podID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerID="9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee" exitCode=0 Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.568515 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" event={"ID":"69fa4317-bafb-462d-b6aa-ea07437277f1","Type":"ContainerDied","Data":"9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.568567 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" event={"ID":"69fa4317-bafb-462d-b6aa-ea07437277f1","Type":"ContainerDied","Data":"a0cec62e2c7e846902634b014c1304ee38e6c7bc3f234649898d90b9aefa0ac7"} Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.568530 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-tvrln" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.568806 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-vc5cr" podStartSLOduration=2.5687593680000003 podStartE2EDuration="2.568759368s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:46.562903705 +0000 UTC m=+1206.931530038" watchObservedRunningTime="2026-03-12 18:22:46.568759368 +0000 UTC m=+1206.937385691" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.613686 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.650377 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6j58l"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.668232 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-6j58l"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.677087 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69fa4317-bafb-462d-b6aa-ea07437277f1" (UID: "69fa4317-bafb-462d-b6aa-ea07437277f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.691986 4926 scope.go:117] "RemoveContainer" containerID="9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.692538 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69fa4317-bafb-462d-b6aa-ea07437277f1" (UID: "69fa4317-bafb-462d-b6aa-ea07437277f1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.704868 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.704890 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.706262 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-config" (OuterVolumeSpecName: "config") pod "69fa4317-bafb-462d-b6aa-ea07437277f1" (UID: "69fa4317-bafb-462d-b6aa-ea07437277f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.718891 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69fa4317-bafb-462d-b6aa-ea07437277f1" (UID: "69fa4317-bafb-462d-b6aa-ea07437277f1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.719231 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69fa4317-bafb-462d-b6aa-ea07437277f1" (UID: "69fa4317-bafb-462d-b6aa-ea07437277f1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.737224 4926 scope.go:117] "RemoveContainer" containerID="fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.807468 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.807498 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.807510 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69fa4317-bafb-462d-b6aa-ea07437277f1-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.818744 4926 scope.go:117] "RemoveContainer" containerID="9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee" Mar 12 18:22:46 crc kubenswrapper[4926]: E0312 18:22:46.826028 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee\": container with ID starting with 9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee not found: ID does not exist" containerID="9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.826080 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee"} err="failed to get container status \"9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee\": rpc error: code = NotFound desc = could not find container \"9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee\": container with ID starting with 9335e26e5334fd57f90ea438fbbb564ec2611ad75fe24dbee8759015a015abee not found: ID does not exist" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.826114 4926 scope.go:117] "RemoveContainer" containerID="fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b" Mar 12 18:22:46 crc kubenswrapper[4926]: E0312 18:22:46.826635 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b\": container with ID starting with fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b not found: ID does not exist" containerID="fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.826767 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b"} err="failed to get container status \"fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b\": rpc error: code = NotFound desc = could not find container \"fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b\": container with ID starting with fa78af8d923d704df31a407f7e84a2ed31c097099bad226b3668d09e28bca79b not found: ID does not exist" Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.934032 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tvrln"] Mar 12 18:22:46 crc kubenswrapper[4926]: I0312 18:22:46.953614 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-tvrln"] Mar 12 18:22:47 crc kubenswrapper[4926]: I0312 18:22:47.280897 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-749fd9494c-gtzjf"] Mar 12 18:22:47 crc kubenswrapper[4926]: W0312 18:22:47.283201 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04df6e00_4552_471c_9ae0_f45362e7e2b4.slice/crio-2d38489365b71e942082167910ea77ad7b51873f3aed248a2bfa421546b7758e WatchSource:0}: Error finding container 2d38489365b71e942082167910ea77ad7b51873f3aed248a2bfa421546b7758e: Status 404 returned error can't find the container with id 2d38489365b71e942082167910ea77ad7b51873f3aed248a2bfa421546b7758e Mar 12 18:22:47 crc kubenswrapper[4926]: I0312 18:22:47.598988 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2900c0f0-0ed1-493a-8bad-ed203682251d","Type":"ContainerStarted","Data":"e192a80af72131c894390e82b963179cefd9639c84c7a61aeb7db9b8d5176057"} Mar 12 18:22:47 crc kubenswrapper[4926]: I0312 18:22:47.610879 4926 generic.go:334] "Generic (PLEG): container finished" podID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerID="d817244fbaa6cac20d7a51e9c98e06316cbeae770e3def99741dd49bb46b1970" exitCode=0 Mar 12 18:22:47 crc kubenswrapper[4926]: I0312 18:22:47.611078 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" event={"ID":"af5b15ef-fa73-4ef4-9235-41da81503d2c","Type":"ContainerDied","Data":"d817244fbaa6cac20d7a51e9c98e06316cbeae770e3def99741dd49bb46b1970"} Mar 12 18:22:47 crc kubenswrapper[4926]: I0312 18:22:47.619713 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-749fd9494c-gtzjf" event={"ID":"04df6e00-4552-471c-9ae0-f45362e7e2b4","Type":"ContainerStarted","Data":"2d38489365b71e942082167910ea77ad7b51873f3aed248a2bfa421546b7758e"} Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.503843 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69fa4317-bafb-462d-b6aa-ea07437277f1" path="/var/lib/kubelet/pods/69fa4317-bafb-462d-b6aa-ea07437277f1/volumes" Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.505071 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c72142af-cd95-4044-92d7-ce173109afa7" path="/var/lib/kubelet/pods/c72142af-cd95-4044-92d7-ce173109afa7/volumes" Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.646895 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2900c0f0-0ed1-493a-8bad-ed203682251d","Type":"ContainerStarted","Data":"e3c316681f99a8d367b4d444d86939314f7068d8029528e7292763d77248f72b"} Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.647052 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-log" containerID="cri-o://e192a80af72131c894390e82b963179cefd9639c84c7a61aeb7db9b8d5176057" gracePeriod=30 Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.647512 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-httpd" containerID="cri-o://e3c316681f99a8d367b4d444d86939314f7068d8029528e7292763d77248f72b" gracePeriod=30 Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.656673 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" event={"ID":"af5b15ef-fa73-4ef4-9235-41da81503d2c","Type":"ContainerStarted","Data":"acee92af6092ee87ba774fbb0a9ef600de19a9fc8c8b93c392fa1761ce8a5481"} Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.657784 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.681623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a40ca3d-194d-4baa-8417-3614c8aeef08","Type":"ContainerStarted","Data":"c5bf2f687495f9063b1f714865d39313a50f9f846c3751dc36bfcd3a2678bb59"} Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.681686 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a40ca3d-194d-4baa-8417-3614c8aeef08","Type":"ContainerStarted","Data":"d6e2011fce2ea9b854a62890216980791595e226fd3037e1e5d5033caf9aa264"} Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.681832 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-log" containerID="cri-o://d6e2011fce2ea9b854a62890216980791595e226fd3037e1e5d5033caf9aa264" gracePeriod=30 Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.682129 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-httpd" containerID="cri-o://c5bf2f687495f9063b1f714865d39313a50f9f846c3751dc36bfcd3a2678bb59" gracePeriod=30 Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.687118 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.687096939 podStartE2EDuration="4.687096939s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:48.665739305 +0000 UTC m=+1209.034365648" watchObservedRunningTime="2026-03-12 18:22:48.687096939 +0000 UTC m=+1209.055723272" Mar 12 18:22:48 crc kubenswrapper[4926]: I0312 18:22:48.695969 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" podStartSLOduration=4.695951354 podStartE2EDuration="4.695951354s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:48.687340587 +0000 UTC m=+1209.055966920" watchObservedRunningTime="2026-03-12 18:22:48.695951354 +0000 UTC m=+1209.064577677" Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.694772 4926 generic.go:334] "Generic (PLEG): container finished" podID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerID="e3c316681f99a8d367b4d444d86939314f7068d8029528e7292763d77248f72b" exitCode=0 Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.695095 4926 generic.go:334] "Generic (PLEG): container finished" podID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerID="e192a80af72131c894390e82b963179cefd9639c84c7a61aeb7db9b8d5176057" exitCode=143 Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.695133 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2900c0f0-0ed1-493a-8bad-ed203682251d","Type":"ContainerDied","Data":"e3c316681f99a8d367b4d444d86939314f7068d8029528e7292763d77248f72b"} Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.695158 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2900c0f0-0ed1-493a-8bad-ed203682251d","Type":"ContainerDied","Data":"e192a80af72131c894390e82b963179cefd9639c84c7a61aeb7db9b8d5176057"} Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.697702 4926 generic.go:334] "Generic (PLEG): container finished" podID="fdce3e2c-d108-4904-b4ab-8d95d4838bad" containerID="334e13044642c4b0a613f2bfccdc8e4e0067f4f734093e62a567052168539e3c" exitCode=0 Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.697773 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hrv24" event={"ID":"fdce3e2c-d108-4904-b4ab-8d95d4838bad","Type":"ContainerDied","Data":"334e13044642c4b0a613f2bfccdc8e4e0067f4f734093e62a567052168539e3c"} Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.701049 4926 generic.go:334] "Generic (PLEG): container finished" podID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerID="c5bf2f687495f9063b1f714865d39313a50f9f846c3751dc36bfcd3a2678bb59" exitCode=143 Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.701071 4926 generic.go:334] "Generic (PLEG): container finished" podID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerID="d6e2011fce2ea9b854a62890216980791595e226fd3037e1e5d5033caf9aa264" exitCode=143 Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.701137 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a40ca3d-194d-4baa-8417-3614c8aeef08","Type":"ContainerDied","Data":"c5bf2f687495f9063b1f714865d39313a50f9f846c3751dc36bfcd3a2678bb59"} Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.701194 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a40ca3d-194d-4baa-8417-3614c8aeef08","Type":"ContainerDied","Data":"d6e2011fce2ea9b854a62890216980791595e226fd3037e1e5d5033caf9aa264"} Mar 12 18:22:49 crc kubenswrapper[4926]: I0312 18:22:49.720619 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.720599672 podStartE2EDuration="5.720599672s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:22:48.719910269 +0000 UTC m=+1209.088536642" watchObservedRunningTime="2026-03-12 18:22:49.720599672 +0000 UTC m=+1210.089226005" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.858136 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84786596b5-28lbk"] Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.882138 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-89554fb64-s9c6q"] Mar 12 18:22:52 crc kubenswrapper[4926]: E0312 18:22:52.882481 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerName="dnsmasq-dns" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.882494 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerName="dnsmasq-dns" Mar 12 18:22:52 crc kubenswrapper[4926]: E0312 18:22:52.882509 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerName="init" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.882515 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerName="init" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.882675 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="69fa4317-bafb-462d-b6aa-ea07437277f1" containerName="dnsmasq-dns" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.883534 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.888759 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.905207 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-89554fb64-s9c6q"] Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.958578 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p78br\" (UniqueName: \"kubernetes.io/projected/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-kube-api-access-p78br\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.958644 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-scripts\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.958668 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-config-data\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.958698 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-secret-key\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.958714 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-logs\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.958729 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-combined-ca-bundle\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.958830 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-tls-certs\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:52 crc kubenswrapper[4926]: I0312 18:22:52.965978 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-749fd9494c-gtzjf"] Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.000778 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5c6848d8cd-cq57n"] Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.002286 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.036000 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c6848d8cd-cq57n"] Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.060892 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-horizon-tls-certs\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.060945 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-scripts\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.060994 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-tls-certs\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061035 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmcq8\" (UniqueName: \"kubernetes.io/projected/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-kube-api-access-xmcq8\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061053 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-config-data\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061073 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p78br\" (UniqueName: \"kubernetes.io/projected/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-kube-api-access-p78br\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061102 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-combined-ca-bundle\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061127 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-logs\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061144 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-scripts\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061161 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-horizon-secret-key\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-config-data\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061209 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-secret-key\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061228 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-logs\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.061249 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-combined-ca-bundle\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.062233 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-logs\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.062654 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-scripts\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.063231 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-config-data\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.070695 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-combined-ca-bundle\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.072261 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-secret-key\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.077242 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-tls-certs\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.082590 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p78br\" (UniqueName: \"kubernetes.io/projected/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-kube-api-access-p78br\") pod \"horizon-89554fb64-s9c6q\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.162415 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmcq8\" (UniqueName: \"kubernetes.io/projected/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-kube-api-access-xmcq8\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.162962 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-config-data\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.163030 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-combined-ca-bundle\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.163092 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-logs\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.163112 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-horizon-secret-key\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.163247 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-horizon-tls-certs\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.163282 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-scripts\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.164209 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-logs\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.164277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-config-data\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.164462 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-scripts\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.178021 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-combined-ca-bundle\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.178535 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-horizon-tls-certs\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.179820 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-horizon-secret-key\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.181253 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmcq8\" (UniqueName: \"kubernetes.io/projected/a1ae8f23-3518-430a-bbcf-e7be0cb8282e-kube-api-access-xmcq8\") pod \"horizon-5c6848d8cd-cq57n\" (UID: \"a1ae8f23-3518-430a-bbcf-e7be0cb8282e\") " pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.216735 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:22:53 crc kubenswrapper[4926]: I0312 18:22:53.317499 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:22:55 crc kubenswrapper[4926]: I0312 18:22:55.070623 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:22:55 crc kubenswrapper[4926]: I0312 18:22:55.145349 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7vm4l"] Mar 12 18:22:55 crc kubenswrapper[4926]: I0312 18:22:55.145640 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-7vm4l" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="dnsmasq-dns" containerID="cri-o://633289debf3f6b1a2f62e3d378805f919d6393fc63c3e3559b1339530737acbe" gracePeriod=10 Mar 12 18:22:55 crc kubenswrapper[4926]: I0312 18:22:55.757147 4926 generic.go:334] "Generic (PLEG): container finished" podID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerID="633289debf3f6b1a2f62e3d378805f919d6393fc63c3e3559b1339530737acbe" exitCode=0 Mar 12 18:22:55 crc kubenswrapper[4926]: I0312 18:22:55.757323 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7vm4l" event={"ID":"f569f4cb-b487-41d8-bab4-5c2d7aba2219","Type":"ContainerDied","Data":"633289debf3f6b1a2f62e3d378805f919d6393fc63c3e3559b1339530737acbe"} Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.682988 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.698270 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771579 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-httpd-run\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771638 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-public-tls-certs\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771697 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-combined-ca-bundle\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-config-data\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771783 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-config-data\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771828 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-logs\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-internal-tls-certs\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771903 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xn68d\" (UniqueName: \"kubernetes.io/projected/7a40ca3d-194d-4baa-8417-3614c8aeef08-kube-api-access-xn68d\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.771971 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772002 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-logs\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772054 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2njd\" (UniqueName: \"kubernetes.io/projected/2900c0f0-0ed1-493a-8bad-ed203682251d-kube-api-access-z2njd\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772080 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-scripts\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772110 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-combined-ca-bundle\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772133 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-scripts\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772160 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"2900c0f0-0ed1-493a-8bad-ed203682251d\" (UID: \"2900c0f0-0ed1-493a-8bad-ed203682251d\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772220 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-httpd-run\") pod \"7a40ca3d-194d-4baa-8417-3614c8aeef08\" (UID: \"7a40ca3d-194d-4baa-8417-3614c8aeef08\") " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772306 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772619 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-logs" (OuterVolumeSpecName: "logs") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772863 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772871 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.772897 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.778580 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2900c0f0-0ed1-493a-8bad-ed203682251d-kube-api-access-z2njd" (OuterVolumeSpecName: "kube-api-access-z2njd") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "kube-api-access-z2njd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.778888 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-logs" (OuterVolumeSpecName: "logs") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.781793 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-scripts" (OuterVolumeSpecName: "scripts") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.781833 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.783724 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-scripts" (OuterVolumeSpecName: "scripts") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.786052 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a40ca3d-194d-4baa-8417-3614c8aeef08","Type":"ContainerDied","Data":"d093554ce00472559ced5d4b721fe022998a063b108c7d7113b402e4a98bca54"} Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.786080 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.786180 4926 scope.go:117] "RemoveContainer" containerID="c5bf2f687495f9063b1f714865d39313a50f9f846c3751dc36bfcd3a2678bb59" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.787638 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a40ca3d-194d-4baa-8417-3614c8aeef08-kube-api-access-xn68d" (OuterVolumeSpecName: "kube-api-access-xn68d") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "kube-api-access-xn68d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.795044 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2900c0f0-0ed1-493a-8bad-ed203682251d","Type":"ContainerDied","Data":"7a86c1cac56bc2792ca8a01485262ffdadcde243619bf97536386064fe81ed59"} Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.795127 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.801021 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.812195 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.813699 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.832619 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-config-data" (OuterVolumeSpecName: "config-data") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.833470 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.833525 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-config-data" (OuterVolumeSpecName: "config-data") pod "7a40ca3d-194d-4baa-8417-3614c8aeef08" (UID: "7a40ca3d-194d-4baa-8417-3614c8aeef08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.834004 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2900c0f0-0ed1-493a-8bad-ed203682251d" (UID: "2900c0f0-0ed1-493a-8bad-ed203682251d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875206 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875239 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875252 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xn68d\" (UniqueName: \"kubernetes.io/projected/7a40ca3d-194d-4baa-8417-3614c8aeef08-kube-api-access-xn68d\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875285 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875296 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2900c0f0-0ed1-493a-8bad-ed203682251d-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875307 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2njd\" (UniqueName: \"kubernetes.io/projected/2900c0f0-0ed1-493a-8bad-ed203682251d-kube-api-access-z2njd\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875315 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875324 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875332 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875343 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875351 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a40ca3d-194d-4baa-8417-3614c8aeef08-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875360 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875368 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2900c0f0-0ed1-493a-8bad-ed203682251d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.875376 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a40ca3d-194d-4baa-8417-3614c8aeef08-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.894525 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.895966 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.976421 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:57 crc kubenswrapper[4926]: I0312 18:22:57.976477 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.019834 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-7vm4l" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.135185 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.162009 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.182770 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.203515 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.209664 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: E0312 18:22:58.209968 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-log" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.209984 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-log" Mar 12 18:22:58 crc kubenswrapper[4926]: E0312 18:22:58.209995 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-log" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.210003 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-log" Mar 12 18:22:58 crc kubenswrapper[4926]: E0312 18:22:58.210023 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-httpd" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.210031 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-httpd" Mar 12 18:22:58 crc kubenswrapper[4926]: E0312 18:22:58.210041 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-httpd" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.210047 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-httpd" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.210204 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-log" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.210218 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-httpd" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.210227 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" containerName="glance-httpd" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.210244 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" containerName="glance-log" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.213219 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.227400 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.227779 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.237838 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2zfbp" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.238050 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.256917 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.258466 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.264493 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.265126 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.271257 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281535 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281595 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qgwd\" (UniqueName: \"kubernetes.io/projected/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-kube-api-access-7qgwd\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281642 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281687 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281725 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281751 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281786 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.281827 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.291906 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384721 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384772 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384801 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384834 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qgwd\" (UniqueName: \"kubernetes.io/projected/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-kube-api-access-7qgwd\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384876 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384924 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384957 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.384982 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385015 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385067 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-config-data\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385108 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385151 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-scripts\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v2m5\" (UniqueName: \"kubernetes.io/projected/27561f0e-1da4-4313-a7df-544fdfc893b1-kube-api-access-9v2m5\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385217 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-logs\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.385250 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.387981 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.392522 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.393348 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.395779 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.396248 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.398274 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.399291 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.424602 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qgwd\" (UniqueName: \"kubernetes.io/projected/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-kube-api-access-7qgwd\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.431373 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.486837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.486904 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.486923 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.487004 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.487023 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-config-data\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.487061 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-scripts\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.487081 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v2m5\" (UniqueName: \"kubernetes.io/projected/27561f0e-1da4-4313-a7df-544fdfc893b1-kube-api-access-9v2m5\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.487132 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-logs\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.487234 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.487606 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-logs\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.488967 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.493943 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.499001 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-config-data\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.499683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.501165 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2900c0f0-0ed1-493a-8bad-ed203682251d" path="/var/lib/kubelet/pods/2900c0f0-0ed1-493a-8bad-ed203682251d/volumes" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.502859 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a40ca3d-194d-4baa-8417-3614c8aeef08" path="/var/lib/kubelet/pods/7a40ca3d-194d-4baa-8417-3614c8aeef08/volumes" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.505635 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-scripts\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.509951 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v2m5\" (UniqueName: \"kubernetes.io/projected/27561f0e-1da4-4313-a7df-544fdfc893b1-kube-api-access-9v2m5\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.518940 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " pod="openstack/glance-default-external-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.604030 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:22:58 crc kubenswrapper[4926]: I0312 18:22:58.656886 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:23:02 crc kubenswrapper[4926]: I0312 18:23:02.836144 4926 scope.go:117] "RemoveContainer" containerID="d82acd09ec6775228cff799be28fec22c80e7a284a7159b9a2b0821199a8fa17" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.003864 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.004051 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5c5h548hbdhf8h556h5c5hd4h5c8h87h5ffh7bh59h694hf7h5ddh7bh5bhdfhdch657hd4h654h68fh65bh5b5h64h658h66dh577h8dhc8h5cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-flm2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7786dc65fc-n6hp6_openstack(8c2a8323-527c-4e19-ab55-3c291c4d538f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.006349 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7786dc65fc-n6hp6" podUID="8c2a8323-527c-4e19-ab55-3c291c4d538f" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.019772 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-7vm4l" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.020103 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.020302 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nf4h5c4hcdh567h54ch555hdh97h696h645h89hc6h5bfh64h5cchdfh544hch596hf9h547h699h676h59bh89h55dh676h675h58bh555h5fch697q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8g599,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-749fd9494c-gtzjf_openstack(04df6e00-4552-471c-9ae0-f45362e7e2b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.026095 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.026227 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5f5hb6h5bh564h58fh5b7h65ch689h644h5f6h8ch58dh679h599h7fh555h586h5d8h57dh5b9h66fh5d7h59fh8h678h9h6chb9h7ch58bh5d5h66bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p7bxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-84786596b5-28lbk_openstack(1ca9ea6f-908c-48db-9313-c3ff4809a993): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.027607 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-749fd9494c-gtzjf" podUID="04df6e00-4552-471c-9ae0-f45362e7e2b4" Mar 12 18:23:03 crc kubenswrapper[4926]: E0312 18:23:03.028396 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-84786596b5-28lbk" podUID="1ca9ea6f-908c-48db-9313-c3ff4809a993" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.083407 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.106050 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-combined-ca-bundle\") pod \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.106112 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpnk7\" (UniqueName: \"kubernetes.io/projected/fdce3e2c-d108-4904-b4ab-8d95d4838bad-kube-api-access-rpnk7\") pod \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.106164 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-scripts\") pod \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.106233 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-config-data\") pod \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.106288 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-credential-keys\") pod \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.106361 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-fernet-keys\") pod \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\" (UID: \"fdce3e2c-d108-4904-b4ab-8d95d4838bad\") " Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.113708 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "fdce3e2c-d108-4904-b4ab-8d95d4838bad" (UID: "fdce3e2c-d108-4904-b4ab-8d95d4838bad"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.114280 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdce3e2c-d108-4904-b4ab-8d95d4838bad-kube-api-access-rpnk7" (OuterVolumeSpecName: "kube-api-access-rpnk7") pod "fdce3e2c-d108-4904-b4ab-8d95d4838bad" (UID: "fdce3e2c-d108-4904-b4ab-8d95d4838bad"). InnerVolumeSpecName "kube-api-access-rpnk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.115285 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-scripts" (OuterVolumeSpecName: "scripts") pod "fdce3e2c-d108-4904-b4ab-8d95d4838bad" (UID: "fdce3e2c-d108-4904-b4ab-8d95d4838bad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.139274 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fdce3e2c-d108-4904-b4ab-8d95d4838bad" (UID: "fdce3e2c-d108-4904-b4ab-8d95d4838bad"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.143996 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-config-data" (OuterVolumeSpecName: "config-data") pod "fdce3e2c-d108-4904-b4ab-8d95d4838bad" (UID: "fdce3e2c-d108-4904-b4ab-8d95d4838bad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.150550 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdce3e2c-d108-4904-b4ab-8d95d4838bad" (UID: "fdce3e2c-d108-4904-b4ab-8d95d4838bad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.208037 4926 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.208065 4926 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.208075 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.208084 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpnk7\" (UniqueName: \"kubernetes.io/projected/fdce3e2c-d108-4904-b4ab-8d95d4838bad-kube-api-access-rpnk7\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.208094 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.208101 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fdce3e2c-d108-4904-b4ab-8d95d4838bad-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.846168 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hrv24" Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.846029 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hrv24" event={"ID":"fdce3e2c-d108-4904-b4ab-8d95d4838bad","Type":"ContainerDied","Data":"cbd223739736371712c30899d1265bcdf31be415ef718a7391e85bb96801c5f5"} Mar 12 18:23:03 crc kubenswrapper[4926]: I0312 18:23:03.861091 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbd223739736371712c30899d1265bcdf31be415ef718a7391e85bb96801c5f5" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.191736 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hrv24"] Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.193178 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hrv24"] Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.263724 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-98lfj"] Mar 12 18:23:04 crc kubenswrapper[4926]: E0312 18:23:04.264167 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdce3e2c-d108-4904-b4ab-8d95d4838bad" containerName="keystone-bootstrap" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.264190 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdce3e2c-d108-4904-b4ab-8d95d4838bad" containerName="keystone-bootstrap" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.264360 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdce3e2c-d108-4904-b4ab-8d95d4838bad" containerName="keystone-bootstrap" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.266353 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.269995 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vrsc7" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.270154 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.270267 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.270374 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.270913 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.281252 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-98lfj"] Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.335000 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-scripts\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.335052 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-fernet-keys\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.335076 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbwc8\" (UniqueName: \"kubernetes.io/projected/af5704dd-cd13-4e5f-a77b-01266c63eeba-kube-api-access-zbwc8\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.335143 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-credential-keys\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.335203 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-combined-ca-bundle\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.335262 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-config-data\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.436286 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-credential-keys\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.436349 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-combined-ca-bundle\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.436396 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-config-data\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.436453 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-scripts\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.436477 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-fernet-keys\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.436493 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbwc8\" (UniqueName: \"kubernetes.io/projected/af5704dd-cd13-4e5f-a77b-01266c63eeba-kube-api-access-zbwc8\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.441548 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-combined-ca-bundle\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.441616 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-config-data\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.441880 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-fernet-keys\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.442237 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-scripts\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.446766 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-credential-keys\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.451049 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbwc8\" (UniqueName: \"kubernetes.io/projected/af5704dd-cd13-4e5f-a77b-01266c63eeba-kube-api-access-zbwc8\") pod \"keystone-bootstrap-98lfj\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.499618 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdce3e2c-d108-4904-b4ab-8d95d4838bad" path="/var/lib/kubelet/pods/fdce3e2c-d108-4904-b4ab-8d95d4838bad/volumes" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.601004 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.857499 4926 generic.go:334] "Generic (PLEG): container finished" podID="e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" containerID="d40939b357197123b7378525a59c3f91e7240cb30140873883a85a0672ad25de" exitCode=0 Mar 12 18:23:04 crc kubenswrapper[4926]: I0312 18:23:04.857540 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vc5cr" event={"ID":"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b","Type":"ContainerDied","Data":"d40939b357197123b7378525a59c3f91e7240cb30140873883a85a0672ad25de"} Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.594694 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.602088 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.610348 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.615168 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.637630 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.750320 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-config\") pod \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.750394 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ca9ea6f-908c-48db-9313-c3ff4809a993-horizon-secret-key\") pod \"1ca9ea6f-908c-48db-9313-c3ff4809a993\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.750421 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vv46v\" (UniqueName: \"kubernetes.io/projected/f569f4cb-b487-41d8-bab4-5c2d7aba2219-kube-api-access-vv46v\") pod \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.750500 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df6kk\" (UniqueName: \"kubernetes.io/projected/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-kube-api-access-df6kk\") pod \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.750534 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-config-data\") pod \"8c2a8323-527c-4e19-ab55-3c291c4d538f\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.750610 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-combined-ca-bundle\") pod \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\" (UID: \"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.751391 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-scripts\") pod \"1ca9ea6f-908c-48db-9313-c3ff4809a993\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.751471 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-sb\") pod \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.751505 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-nb\") pod \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.751576 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-scripts\") pod \"04df6e00-4552-471c-9ae0-f45362e7e2b4\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.751742 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7bxq\" (UniqueName: \"kubernetes.io/projected/1ca9ea6f-908c-48db-9313-c3ff4809a993-kube-api-access-p7bxq\") pod \"1ca9ea6f-908c-48db-9313-c3ff4809a993\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.751809 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ca9ea6f-908c-48db-9313-c3ff4809a993-logs\") pod \"1ca9ea6f-908c-48db-9313-c3ff4809a993\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.751842 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-config-data\") pod \"1ca9ea6f-908c-48db-9313-c3ff4809a993\" (UID: \"1ca9ea6f-908c-48db-9313-c3ff4809a993\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752081 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-config-data" (OuterVolumeSpecName: "config-data") pod "8c2a8323-527c-4e19-ab55-3c291c4d538f" (UID: "8c2a8323-527c-4e19-ab55-3c291c4d538f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752111 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-dns-svc\") pod \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752188 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2a8323-527c-4e19-ab55-3c291c4d538f-logs\") pod \"8c2a8323-527c-4e19-ab55-3c291c4d538f\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752287 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-config\") pod \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\" (UID: \"f569f4cb-b487-41d8-bab4-5c2d7aba2219\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752331 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flm2j\" (UniqueName: \"kubernetes.io/projected/8c2a8323-527c-4e19-ab55-3c291c4d538f-kube-api-access-flm2j\") pod \"8c2a8323-527c-4e19-ab55-3c291c4d538f\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752355 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04df6e00-4552-471c-9ae0-f45362e7e2b4-logs\") pod \"04df6e00-4552-471c-9ae0-f45362e7e2b4\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752375 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g599\" (UniqueName: \"kubernetes.io/projected/04df6e00-4552-471c-9ae0-f45362e7e2b4-kube-api-access-8g599\") pod \"04df6e00-4552-471c-9ae0-f45362e7e2b4\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752656 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-scripts" (OuterVolumeSpecName: "scripts") pod "04df6e00-4552-471c-9ae0-f45362e7e2b4" (UID: "04df6e00-4552-471c-9ae0-f45362e7e2b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752755 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c2a8323-527c-4e19-ab55-3c291c4d538f-horizon-secret-key\") pod \"8c2a8323-527c-4e19-ab55-3c291c4d538f\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752795 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04df6e00-4552-471c-9ae0-f45362e7e2b4-horizon-secret-key\") pod \"04df6e00-4552-471c-9ae0-f45362e7e2b4\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752828 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-config-data\") pod \"04df6e00-4552-471c-9ae0-f45362e7e2b4\" (UID: \"04df6e00-4552-471c-9ae0-f45362e7e2b4\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.752852 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-scripts\") pod \"8c2a8323-527c-4e19-ab55-3c291c4d538f\" (UID: \"8c2a8323-527c-4e19-ab55-3c291c4d538f\") " Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.753083 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ca9ea6f-908c-48db-9313-c3ff4809a993-logs" (OuterVolumeSpecName: "logs") pod "1ca9ea6f-908c-48db-9313-c3ff4809a993" (UID: "1ca9ea6f-908c-48db-9313-c3ff4809a993"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.753543 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-scripts" (OuterVolumeSpecName: "scripts") pod "8c2a8323-527c-4e19-ab55-3c291c4d538f" (UID: "8c2a8323-527c-4e19-ab55-3c291c4d538f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.753553 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-scripts" (OuterVolumeSpecName: "scripts") pod "1ca9ea6f-908c-48db-9313-c3ff4809a993" (UID: "1ca9ea6f-908c-48db-9313-c3ff4809a993"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.753865 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.754152 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.754239 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ca9ea6f-908c-48db-9313-c3ff4809a993-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.754315 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8c2a8323-527c-4e19-ab55-3c291c4d538f-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.755822 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-kube-api-access-df6kk" (OuterVolumeSpecName: "kube-api-access-df6kk") pod "e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" (UID: "e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b"). InnerVolumeSpecName "kube-api-access-df6kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.756087 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c2a8323-527c-4e19-ab55-3c291c4d538f-logs" (OuterVolumeSpecName: "logs") pod "8c2a8323-527c-4e19-ab55-3c291c4d538f" (UID: "8c2a8323-527c-4e19-ab55-3c291c4d538f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.756692 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca9ea6f-908c-48db-9313-c3ff4809a993-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1ca9ea6f-908c-48db-9313-c3ff4809a993" (UID: "1ca9ea6f-908c-48db-9313-c3ff4809a993"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.757284 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c2a8323-527c-4e19-ab55-3c291c4d538f-kube-api-access-flm2j" (OuterVolumeSpecName: "kube-api-access-flm2j") pod "8c2a8323-527c-4e19-ab55-3c291c4d538f" (UID: "8c2a8323-527c-4e19-ab55-3c291c4d538f"). InnerVolumeSpecName "kube-api-access-flm2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.757563 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04df6e00-4552-471c-9ae0-f45362e7e2b4-logs" (OuterVolumeSpecName: "logs") pod "04df6e00-4552-471c-9ae0-f45362e7e2b4" (UID: "04df6e00-4552-471c-9ae0-f45362e7e2b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.758035 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-config-data" (OuterVolumeSpecName: "config-data") pod "1ca9ea6f-908c-48db-9313-c3ff4809a993" (UID: "1ca9ea6f-908c-48db-9313-c3ff4809a993"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.760332 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-config-data" (OuterVolumeSpecName: "config-data") pod "04df6e00-4552-471c-9ae0-f45362e7e2b4" (UID: "04df6e00-4552-471c-9ae0-f45362e7e2b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.761075 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04df6e00-4552-471c-9ae0-f45362e7e2b4-kube-api-access-8g599" (OuterVolumeSpecName: "kube-api-access-8g599") pod "04df6e00-4552-471c-9ae0-f45362e7e2b4" (UID: "04df6e00-4552-471c-9ae0-f45362e7e2b4"). InnerVolumeSpecName "kube-api-access-8g599". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.761621 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2a8323-527c-4e19-ab55-3c291c4d538f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8c2a8323-527c-4e19-ab55-3c291c4d538f" (UID: "8c2a8323-527c-4e19-ab55-3c291c4d538f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.764191 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f569f4cb-b487-41d8-bab4-5c2d7aba2219-kube-api-access-vv46v" (OuterVolumeSpecName: "kube-api-access-vv46v") pod "f569f4cb-b487-41d8-bab4-5c2d7aba2219" (UID: "f569f4cb-b487-41d8-bab4-5c2d7aba2219"). InnerVolumeSpecName "kube-api-access-vv46v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.764981 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04df6e00-4552-471c-9ae0-f45362e7e2b4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "04df6e00-4552-471c-9ae0-f45362e7e2b4" (UID: "04df6e00-4552-471c-9ae0-f45362e7e2b4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.780605 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca9ea6f-908c-48db-9313-c3ff4809a993-kube-api-access-p7bxq" (OuterVolumeSpecName: "kube-api-access-p7bxq") pod "1ca9ea6f-908c-48db-9313-c3ff4809a993" (UID: "1ca9ea6f-908c-48db-9313-c3ff4809a993"). InnerVolumeSpecName "kube-api-access-p7bxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.790524 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-config" (OuterVolumeSpecName: "config") pod "e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" (UID: "e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.792894 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" (UID: "e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.805475 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f569f4cb-b487-41d8-bab4-5c2d7aba2219" (UID: "f569f4cb-b487-41d8-bab4-5c2d7aba2219"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.812030 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-config" (OuterVolumeSpecName: "config") pod "f569f4cb-b487-41d8-bab4-5c2d7aba2219" (UID: "f569f4cb-b487-41d8-bab4-5c2d7aba2219"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.814147 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f569f4cb-b487-41d8-bab4-5c2d7aba2219" (UID: "f569f4cb-b487-41d8-bab4-5c2d7aba2219"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.817670 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f569f4cb-b487-41d8-bab4-5c2d7aba2219" (UID: "f569f4cb-b487-41d8-bab4-5c2d7aba2219"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860675 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df6kk\" (UniqueName: \"kubernetes.io/projected/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-kube-api-access-df6kk\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860721 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860730 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860739 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860749 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860760 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7bxq\" (UniqueName: \"kubernetes.io/projected/1ca9ea6f-908c-48db-9313-c3ff4809a993-kube-api-access-p7bxq\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860768 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ca9ea6f-908c-48db-9313-c3ff4809a993-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860776 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860784 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c2a8323-527c-4e19-ab55-3c291c4d538f-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860792 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f569f4cb-b487-41d8-bab4-5c2d7aba2219-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860800 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flm2j\" (UniqueName: \"kubernetes.io/projected/8c2a8323-527c-4e19-ab55-3c291c4d538f-kube-api-access-flm2j\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860808 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04df6e00-4552-471c-9ae0-f45362e7e2b4-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860817 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g599\" (UniqueName: \"kubernetes.io/projected/04df6e00-4552-471c-9ae0-f45362e7e2b4-kube-api-access-8g599\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860826 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/04df6e00-4552-471c-9ae0-f45362e7e2b4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860834 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8c2a8323-527c-4e19-ab55-3c291c4d538f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860842 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/04df6e00-4552-471c-9ae0-f45362e7e2b4-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860850 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860858 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ca9ea6f-908c-48db-9313-c3ff4809a993-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.860866 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vv46v\" (UniqueName: \"kubernetes.io/projected/f569f4cb-b487-41d8-bab4-5c2d7aba2219-kube-api-access-vv46v\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.905971 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-84786596b5-28lbk" event={"ID":"1ca9ea6f-908c-48db-9313-c3ff4809a993","Type":"ContainerDied","Data":"87827b8a20d43f6415fa4987b8da56ced353164794c7d29438fb6ffa36394a39"} Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.906223 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-84786596b5-28lbk" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.907290 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7786dc65fc-n6hp6" event={"ID":"8c2a8323-527c-4e19-ab55-3c291c4d538f","Type":"ContainerDied","Data":"1a2407e300165d17a9ee94f4b0ab357bc8d93ae3d9e96138519ce22a7544a84a"} Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.907416 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7786dc65fc-n6hp6" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.915394 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-7vm4l" event={"ID":"f569f4cb-b487-41d8-bab4-5c2d7aba2219","Type":"ContainerDied","Data":"1a2ca21647a28c29a2f5caa87bf2be645c7eb58039451905ac1f8540e674ede0"} Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.915427 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-7vm4l" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.917394 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-749fd9494c-gtzjf" event={"ID":"04df6e00-4552-471c-9ae0-f45362e7e2b4","Type":"ContainerDied","Data":"2d38489365b71e942082167910ea77ad7b51873f3aed248a2bfa421546b7758e"} Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.917462 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-749fd9494c-gtzjf" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.920923 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vc5cr" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.920927 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vc5cr" event={"ID":"e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b","Type":"ContainerDied","Data":"85be7b1fc6b389d0d4f2769e75e7678715b75c3e095cab19a3e0aad9bb5656b7"} Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.920966 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85be7b1fc6b389d0d4f2769e75e7678715b75c3e095cab19a3e0aad9bb5656b7" Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.968314 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7786dc65fc-n6hp6"] Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.984255 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7786dc65fc-n6hp6"] Mar 12 18:23:10 crc kubenswrapper[4926]: I0312 18:23:10.997929 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-84786596b5-28lbk"] Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.005494 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-84786596b5-28lbk"] Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.014598 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7vm4l"] Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.017930 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-7vm4l"] Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.057821 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-749fd9494c-gtzjf"] Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.068015 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-749fd9494c-gtzjf"] Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.747719 4926 scope.go:117] "RemoveContainer" containerID="d6e2011fce2ea9b854a62890216980791595e226fd3037e1e5d5033caf9aa264" Mar 12 18:23:11 crc kubenswrapper[4926]: E0312 18:23:11.764670 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 12 18:23:11 crc kubenswrapper[4926]: E0312 18:23:11.764954 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fgqfm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-lgvzs_openstack(dac4b5d6-fb31-4955-8679-db9d3ff63c10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 12 18:23:11 crc kubenswrapper[4926]: E0312 18:23:11.766202 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-lgvzs" podUID="dac4b5d6-fb31-4955-8679-db9d3ff63c10" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.886901 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rvkqm"] Mar 12 18:23:11 crc kubenswrapper[4926]: E0312 18:23:11.890873 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="init" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.890992 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="init" Mar 12 18:23:11 crc kubenswrapper[4926]: E0312 18:23:11.891097 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" containerName="neutron-db-sync" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.891191 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" containerName="neutron-db-sync" Mar 12 18:23:11 crc kubenswrapper[4926]: E0312 18:23:11.891314 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="dnsmasq-dns" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.891409 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="dnsmasq-dns" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.891757 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" containerName="neutron-db-sync" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.891873 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="dnsmasq-dns" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.893061 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:11 crc kubenswrapper[4926]: I0312 18:23:11.915503 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rvkqm"] Mar 12 18:23:11 crc kubenswrapper[4926]: E0312 18:23:11.979023 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-lgvzs" podUID="dac4b5d6-fb31-4955-8679-db9d3ff63c10" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.023178 4926 scope.go:117] "RemoveContainer" containerID="e3c316681f99a8d367b4d444d86939314f7068d8029528e7292763d77248f72b" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.061387 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-884f7b65b-tpkzl"] Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.062858 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.069933 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.070270 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.070466 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qpbhw" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.070591 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.077992 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-884f7b65b-tpkzl"] Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.100403 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.100506 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwp6h\" (UniqueName: \"kubernetes.io/projected/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-kube-api-access-rwp6h\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.100614 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.100656 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.100762 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-config\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.100833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.191064 4926 scope.go:117] "RemoveContainer" containerID="e192a80af72131c894390e82b963179cefd9639c84c7a61aeb7db9b8d5176057" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.202959 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.209232 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.209801 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-ovndb-tls-certs\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.209918 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.209983 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwp6h\" (UniqueName: \"kubernetes.io/projected/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-kube-api-access-rwp6h\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210022 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-combined-ca-bundle\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210142 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210160 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-httpd-config\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210209 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vz6v\" (UniqueName: \"kubernetes.io/projected/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-kube-api-access-7vz6v\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210230 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-config\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210322 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-config\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.210968 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.212269 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.213110 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-config\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.214460 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.233334 4926 scope.go:117] "RemoveContainer" containerID="633289debf3f6b1a2f62e3d378805f919d6393fc63c3e3559b1339530737acbe" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.236445 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwp6h\" (UniqueName: \"kubernetes.io/projected/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-kube-api-access-rwp6h\") pod \"dnsmasq-dns-55f844cf75-rvkqm\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.238545 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.289740 4926 scope.go:117] "RemoveContainer" containerID="50524e634e7b1a90622cd86f4a3bdca7cea0ea5d1a412cda03a7f2f2cce56ce2" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.312204 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vz6v\" (UniqueName: \"kubernetes.io/projected/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-kube-api-access-7vz6v\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.312243 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-config\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.313547 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-ovndb-tls-certs\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.313648 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-combined-ca-bundle\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.313758 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-httpd-config\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.316336 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-config\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.316834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-ovndb-tls-certs\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.333878 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-httpd-config\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.335294 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-combined-ca-bundle\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.342034 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vz6v\" (UniqueName: \"kubernetes.io/projected/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-kube-api-access-7vz6v\") pod \"neutron-884f7b65b-tpkzl\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.425432 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.526614 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04df6e00-4552-471c-9ae0-f45362e7e2b4" path="/var/lib/kubelet/pods/04df6e00-4552-471c-9ae0-f45362e7e2b4/volumes" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.527301 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca9ea6f-908c-48db-9313-c3ff4809a993" path="/var/lib/kubelet/pods/1ca9ea6f-908c-48db-9313-c3ff4809a993/volumes" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.528232 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c2a8323-527c-4e19-ab55-3c291c4d538f" path="/var/lib/kubelet/pods/8c2a8323-527c-4e19-ab55-3c291c4d538f/volumes" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.528599 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" path="/var/lib/kubelet/pods/f569f4cb-b487-41d8-bab4-5c2d7aba2219/volumes" Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.706561 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5c6848d8cd-cq57n"] Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.718739 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-89554fb64-s9c6q"] Mar 12 18:23:12 crc kubenswrapper[4926]: W0312 18:23:12.724664 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7d07aa_8c5e_49f3_8d85_4c5e9569c572.slice/crio-47a3a3a3dc751bc7ac0243d80b1ecc826d87bf09a948cad8eeaa84ccc31fd39c WatchSource:0}: Error finding container 47a3a3a3dc751bc7ac0243d80b1ecc826d87bf09a948cad8eeaa84ccc31fd39c: Status 404 returned error can't find the container with id 47a3a3a3dc751bc7ac0243d80b1ecc826d87bf09a948cad8eeaa84ccc31fd39c Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.770772 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-98lfj"] Mar 12 18:23:12 crc kubenswrapper[4926]: W0312 18:23:12.776253 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf5704dd_cd13_4e5f_a77b_01266c63eeba.slice/crio-45ccbdc85c6be068623b090cab069fbadeba37022e035780f13ac08ff104b1bc WatchSource:0}: Error finding container 45ccbdc85c6be068623b090cab069fbadeba37022e035780f13ac08ff104b1bc: Status 404 returned error can't find the container with id 45ccbdc85c6be068623b090cab069fbadeba37022e035780f13ac08ff104b1bc Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.987268 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerStarted","Data":"3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c"} Mar 12 18:23:12 crc kubenswrapper[4926]: I0312 18:23:12.994821 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sgnbr" event={"ID":"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6","Type":"ContainerStarted","Data":"d333e02f3e64bf137ca7da16020fb743166cb48ce41f070ab0d454452957be39"} Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:12.999110 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rvkqm"] Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.000111 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-98lfj" event={"ID":"af5704dd-cd13-4e5f-a77b-01266c63eeba","Type":"ContainerStarted","Data":"45ccbdc85c6be068623b090cab069fbadeba37022e035780f13ac08ff104b1bc"} Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.001019 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-txt96" event={"ID":"7a0f2830-bf50-4195-9dad-d4d2c9529ee9","Type":"ContainerStarted","Data":"5b07bfa58e0c68e62027f67884fd04fdf4f157c10b82889fbed4fe1cd06870d8"} Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.017211 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-sgnbr" podStartSLOduration=4.28196941 podStartE2EDuration="29.017192209s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="2026-03-12 18:22:45.724868182 +0000 UTC m=+1206.093494515" lastFinishedPulling="2026-03-12 18:23:10.460090981 +0000 UTC m=+1230.828717314" observedRunningTime="2026-03-12 18:23:13.017078925 +0000 UTC m=+1233.385705258" watchObservedRunningTime="2026-03-12 18:23:13.017192209 +0000 UTC m=+1233.385818542" Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.022273 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-7vm4l" podUID="f569f4cb-b487-41d8-bab4-5c2d7aba2219" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: i/o timeout" Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.022529 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c6848d8cd-cq57n" event={"ID":"a1ae8f23-3518-430a-bbcf-e7be0cb8282e","Type":"ContainerStarted","Data":"1d58f8674df4594ec9f33db698a8a40eaa76148cc32f01ed65517b4662c964ca"} Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.023829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-89554fb64-s9c6q" event={"ID":"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572","Type":"ContainerStarted","Data":"47a3a3a3dc751bc7ac0243d80b1ecc826d87bf09a948cad8eeaa84ccc31fd39c"} Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.033894 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-txt96" podStartSLOduration=4.356917418 podStartE2EDuration="29.033876433s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="2026-03-12 18:22:45.779577084 +0000 UTC m=+1206.148203417" lastFinishedPulling="2026-03-12 18:23:10.456536069 +0000 UTC m=+1230.825162432" observedRunningTime="2026-03-12 18:23:13.03187266 +0000 UTC m=+1233.400498993" watchObservedRunningTime="2026-03-12 18:23:13.033876433 +0000 UTC m=+1233.402502766" Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.067735 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.207636 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-884f7b65b-tpkzl"] Mar 12 18:23:13 crc kubenswrapper[4926]: W0312 18:23:13.211869 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc8f91a_7f34_4957_b3e4_e5d28b2f43ea.slice/crio-4598f84bb743a934c385d17beee4641ea28d3b613e4146bda41f42d7c8f987e3 WatchSource:0}: Error finding container 4598f84bb743a934c385d17beee4641ea28d3b613e4146bda41f42d7c8f987e3: Status 404 returned error can't find the container with id 4598f84bb743a934c385d17beee4641ea28d3b613e4146bda41f42d7c8f987e3 Mar 12 18:23:13 crc kubenswrapper[4926]: I0312 18:23:13.833118 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.034788 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerID="2f62316dc5afc9b2e89e828a9e035d936c5f1005f1c85a5b2bad886d39f3e886" exitCode=0 Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.034851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" event={"ID":"b5cd3f65-2af4-4a38-ab87-c266452f8c5a","Type":"ContainerDied","Data":"2f62316dc5afc9b2e89e828a9e035d936c5f1005f1c85a5b2bad886d39f3e886"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.034875 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" event={"ID":"b5cd3f65-2af4-4a38-ab87-c266452f8c5a","Type":"ContainerStarted","Data":"a1cf427ff4b305faffde9b50b631a3a735bc0b296fb816c491a872c6626c0d60"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.037942 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-884f7b65b-tpkzl" event={"ID":"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea","Type":"ContainerStarted","Data":"e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.037975 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-884f7b65b-tpkzl" event={"ID":"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea","Type":"ContainerStarted","Data":"4598f84bb743a934c385d17beee4641ea28d3b613e4146bda41f42d7c8f987e3"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.043720 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"27561f0e-1da4-4313-a7df-544fdfc893b1","Type":"ContainerStarted","Data":"9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.043758 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"27561f0e-1da4-4313-a7df-544fdfc893b1","Type":"ContainerStarted","Data":"600cdc83756bdbe2006c6d01ba6698b38db80c94206722c7e9c45bd9b1332b71"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.048272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c6848d8cd-cq57n" event={"ID":"a1ae8f23-3518-430a-bbcf-e7be0cb8282e","Type":"ContainerStarted","Data":"0e660ab765557cad74f028b05adf2a2017733c5b7cad2a31d7ba9218f60c1b1a"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.052271 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-89554fb64-s9c6q" event={"ID":"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572","Type":"ContainerStarted","Data":"a41e698428edf8a4dccc211b474a012677761bd1a52d186e37df8c3d1445ee62"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.062521 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-98lfj" event={"ID":"af5704dd-cd13-4e5f-a77b-01266c63eeba","Type":"ContainerStarted","Data":"1b2fbc6d93b78d1252c9b5b62fe658eafa93cdb86d8acefaa37b868dbfa19698"} Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.319781 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-98lfj" podStartSLOduration=10.319763313 podStartE2EDuration="10.319763313s" podCreationTimestamp="2026-03-12 18:23:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:14.079274556 +0000 UTC m=+1234.447900899" watchObservedRunningTime="2026-03-12 18:23:14.319763313 +0000 UTC m=+1234.688389646" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.331385 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-65c5c86775-mct68"] Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.333348 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.336694 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.337225 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.353536 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65c5c86775-mct68"] Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.492956 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-httpd-config\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.493010 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-public-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.493122 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-config\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.493210 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-ovndb-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.493388 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-combined-ca-bundle\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.493431 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzgvh\" (UniqueName: \"kubernetes.io/projected/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-kube-api-access-nzgvh\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.493474 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-internal-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.600940 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-public-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.601107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-config\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.601243 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-ovndb-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.601434 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-combined-ca-bundle\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.601512 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzgvh\" (UniqueName: \"kubernetes.io/projected/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-kube-api-access-nzgvh\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.601546 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-internal-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.601665 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-httpd-config\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.613464 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-internal-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.622762 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-config\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.624732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-ovndb-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.628277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-httpd-config\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.631425 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-combined-ca-bundle\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.632229 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzgvh\" (UniqueName: \"kubernetes.io/projected/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-kube-api-access-nzgvh\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.635214 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-public-tls-certs\") pod \"neutron-65c5c86775-mct68\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:14 crc kubenswrapper[4926]: I0312 18:23:14.654929 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:15 crc kubenswrapper[4926]: I0312 18:23:15.100549 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-884f7b65b-tpkzl" event={"ID":"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea","Type":"ContainerStarted","Data":"d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529"} Mar 12 18:23:15 crc kubenswrapper[4926]: I0312 18:23:15.100959 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:15 crc kubenswrapper[4926]: I0312 18:23:15.102681 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a","Type":"ContainerStarted","Data":"ac524b6af20000079258195bab5b0ef6d7a0ae5f18a8830578a4f4e1b53a9c85"} Mar 12 18:23:15 crc kubenswrapper[4926]: I0312 18:23:15.123938 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-884f7b65b-tpkzl" podStartSLOduration=3.123920696 podStartE2EDuration="3.123920696s" podCreationTimestamp="2026-03-12 18:23:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:15.121364786 +0000 UTC m=+1235.489991119" watchObservedRunningTime="2026-03-12 18:23:15.123920696 +0000 UTC m=+1235.492547029" Mar 12 18:23:15 crc kubenswrapper[4926]: I0312 18:23:15.324950 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-65c5c86775-mct68"] Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.123430 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5c6848d8cd-cq57n" event={"ID":"a1ae8f23-3518-430a-bbcf-e7be0cb8282e","Type":"ContainerStarted","Data":"100a461384abbd26a4bc41fefac81726e40c151486f6992479e0630f10e31127"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.142674 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a","Type":"ContainerStarted","Data":"73938b8cf4a571a4406931a492f2e033c549c10b7a2bcdf0ca6b902ede59649d"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.142713 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a","Type":"ContainerStarted","Data":"809f7347552b8f928619d96194c02f26e8cc98b9b07e8efde999fc4f99c7437f"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.150372 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5c6848d8cd-cq57n" podStartSLOduration=23.530879487 podStartE2EDuration="24.150350544s" podCreationTimestamp="2026-03-12 18:22:52 +0000 UTC" firstStartedPulling="2026-03-12 18:23:12.733682954 +0000 UTC m=+1233.102309297" lastFinishedPulling="2026-03-12 18:23:13.353154021 +0000 UTC m=+1233.721780354" observedRunningTime="2026-03-12 18:23:16.146633698 +0000 UTC m=+1236.515260041" watchObservedRunningTime="2026-03-12 18:23:16.150350544 +0000 UTC m=+1236.518976877" Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.152740 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c5c86775-mct68" event={"ID":"2beed02e-2edf-4b52-8ea6-ae2dae7502d8","Type":"ContainerStarted","Data":"d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.152782 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c5c86775-mct68" event={"ID":"2beed02e-2edf-4b52-8ea6-ae2dae7502d8","Type":"ContainerStarted","Data":"94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.152793 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c5c86775-mct68" event={"ID":"2beed02e-2edf-4b52-8ea6-ae2dae7502d8","Type":"ContainerStarted","Data":"7b5dc2f13fd79b24bbd53ad5c8577f6fcd9167f1b3fea4e0a8139bb248f187ac"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.153662 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.159581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerStarted","Data":"aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.163701 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-89554fb64-s9c6q" event={"ID":"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572","Type":"ContainerStarted","Data":"2388379aefe6f54987e9387c0ab60e55776d5a81c5e2c5b0f21608b48b6e8fa5"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.170930 4926 generic.go:334] "Generic (PLEG): container finished" podID="7a0f2830-bf50-4195-9dad-d4d2c9529ee9" containerID="5b07bfa58e0c68e62027f67884fd04fdf4f157c10b82889fbed4fe1cd06870d8" exitCode=0 Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.173690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-txt96" event={"ID":"7a0f2830-bf50-4195-9dad-d4d2c9529ee9","Type":"ContainerDied","Data":"5b07bfa58e0c68e62027f67884fd04fdf4f157c10b82889fbed4fe1cd06870d8"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.186636 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=18.186622092 podStartE2EDuration="18.186622092s" podCreationTimestamp="2026-03-12 18:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:16.16296131 +0000 UTC m=+1236.531587663" watchObservedRunningTime="2026-03-12 18:23:16.186622092 +0000 UTC m=+1236.555248425" Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.187225 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" event={"ID":"b5cd3f65-2af4-4a38-ab87-c266452f8c5a","Type":"ContainerStarted","Data":"3d75e203f5f24817d9aa7e9a6cf9f743a4b6a96bd884d103fc15a23a410323b0"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.188118 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.194988 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-65c5c86775-mct68" podStartSLOduration=2.194973494 podStartE2EDuration="2.194973494s" podCreationTimestamp="2026-03-12 18:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:16.185804577 +0000 UTC m=+1236.554430910" watchObservedRunningTime="2026-03-12 18:23:16.194973494 +0000 UTC m=+1236.563599827" Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.201684 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"27561f0e-1da4-4313-a7df-544fdfc893b1","Type":"ContainerStarted","Data":"1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178"} Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.236674 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-89554fb64-s9c6q" podStartSLOduration=23.606097966 podStartE2EDuration="24.236656442s" podCreationTimestamp="2026-03-12 18:22:52 +0000 UTC" firstStartedPulling="2026-03-12 18:23:12.729275665 +0000 UTC m=+1233.097901998" lastFinishedPulling="2026-03-12 18:23:13.359834141 +0000 UTC m=+1233.728460474" observedRunningTime="2026-03-12 18:23:16.234067071 +0000 UTC m=+1236.602693404" watchObservedRunningTime="2026-03-12 18:23:16.236656442 +0000 UTC m=+1236.605282775" Mar 12 18:23:16 crc kubenswrapper[4926]: I0312 18:23:16.289855 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" podStartSLOduration=5.289838821 podStartE2EDuration="5.289838821s" podCreationTimestamp="2026-03-12 18:23:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:16.285626269 +0000 UTC m=+1236.654252622" watchObservedRunningTime="2026-03-12 18:23:16.289838821 +0000 UTC m=+1236.658465154" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.211697 4926 generic.go:334] "Generic (PLEG): container finished" podID="e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" containerID="d333e02f3e64bf137ca7da16020fb743166cb48ce41f070ab0d454452957be39" exitCode=0 Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.211794 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sgnbr" event={"ID":"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6","Type":"ContainerDied","Data":"d333e02f3e64bf137ca7da16020fb743166cb48ce41f070ab0d454452957be39"} Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.213626 4926 generic.go:334] "Generic (PLEG): container finished" podID="af5704dd-cd13-4e5f-a77b-01266c63eeba" containerID="1b2fbc6d93b78d1252c9b5b62fe658eafa93cdb86d8acefaa37b868dbfa19698" exitCode=0 Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.214625 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-98lfj" event={"ID":"af5704dd-cd13-4e5f-a77b-01266c63eeba","Type":"ContainerDied","Data":"1b2fbc6d93b78d1252c9b5b62fe658eafa93cdb86d8acefaa37b868dbfa19698"} Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.238886 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=19.23886802 podStartE2EDuration="19.23886802s" podCreationTimestamp="2026-03-12 18:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:16.309472537 +0000 UTC m=+1236.678098890" watchObservedRunningTime="2026-03-12 18:23:17.23886802 +0000 UTC m=+1237.607494353" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.684613 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-txt96" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.789416 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-combined-ca-bundle\") pod \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.789505 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-scripts\") pod \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.789544 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-logs\") pod \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.789596 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgd7q\" (UniqueName: \"kubernetes.io/projected/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-kube-api-access-tgd7q\") pod \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.789715 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-config-data\") pod \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\" (UID: \"7a0f2830-bf50-4195-9dad-d4d2c9529ee9\") " Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.790402 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-logs" (OuterVolumeSpecName: "logs") pod "7a0f2830-bf50-4195-9dad-d4d2c9529ee9" (UID: "7a0f2830-bf50-4195-9dad-d4d2c9529ee9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.794937 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-kube-api-access-tgd7q" (OuterVolumeSpecName: "kube-api-access-tgd7q") pod "7a0f2830-bf50-4195-9dad-d4d2c9529ee9" (UID: "7a0f2830-bf50-4195-9dad-d4d2c9529ee9"). InnerVolumeSpecName "kube-api-access-tgd7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.795092 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-scripts" (OuterVolumeSpecName: "scripts") pod "7a0f2830-bf50-4195-9dad-d4d2c9529ee9" (UID: "7a0f2830-bf50-4195-9dad-d4d2c9529ee9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.814931 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a0f2830-bf50-4195-9dad-d4d2c9529ee9" (UID: "7a0f2830-bf50-4195-9dad-d4d2c9529ee9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.815809 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-config-data" (OuterVolumeSpecName: "config-data") pod "7a0f2830-bf50-4195-9dad-d4d2c9529ee9" (UID: "7a0f2830-bf50-4195-9dad-d4d2c9529ee9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.891983 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgd7q\" (UniqueName: \"kubernetes.io/projected/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-kube-api-access-tgd7q\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.892025 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.892036 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.892045 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:17 crc kubenswrapper[4926]: I0312 18:23:17.892055 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a0f2830-bf50-4195-9dad-d4d2c9529ee9-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.223931 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-txt96" event={"ID":"7a0f2830-bf50-4195-9dad-d4d2c9529ee9","Type":"ContainerDied","Data":"ce15d4b03307ac2a9a7f3c1d05eba2559878bf699c0483537cc9d838c8197125"} Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.224189 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce15d4b03307ac2a9a7f3c1d05eba2559878bf699c0483537cc9d838c8197125" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.224240 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-txt96" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.424584 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6fc994c476-fv9c9"] Mar 12 18:23:18 crc kubenswrapper[4926]: E0312 18:23:18.424996 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a0f2830-bf50-4195-9dad-d4d2c9529ee9" containerName="placement-db-sync" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.425014 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a0f2830-bf50-4195-9dad-d4d2c9529ee9" containerName="placement-db-sync" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.425263 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a0f2830-bf50-4195-9dad-d4d2c9529ee9" containerName="placement-db-sync" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.426687 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.431549 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.431757 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.431895 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.432039 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.432298 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-9lh8l" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.458210 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fc994c476-fv9c9"] Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.502576 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-config-data\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.502627 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-scripts\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.502685 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-public-tls-certs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.502732 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/450e1ecf-5ae7-48b5-b567-e530e254f673-logs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.502753 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-internal-tls-certs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.502846 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh2nd\" (UniqueName: \"kubernetes.io/projected/450e1ecf-5ae7-48b5-b567-e530e254f673-kube-api-access-vh2nd\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.502982 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-combined-ca-bundle\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.608967 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/450e1ecf-5ae7-48b5-b567-e530e254f673-logs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.609034 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-internal-tls-certs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.609073 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh2nd\" (UniqueName: \"kubernetes.io/projected/450e1ecf-5ae7-48b5-b567-e530e254f673-kube-api-access-vh2nd\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.609130 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-combined-ca-bundle\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.609220 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-config-data\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.609240 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-scripts\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.609292 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-public-tls-certs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.610757 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.611586 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/450e1ecf-5ae7-48b5-b567-e530e254f673-logs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.611914 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.619113 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-combined-ca-bundle\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.619645 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-public-tls-certs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.621382 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-scripts\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.627804 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-config-data\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.635516 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh2nd\" (UniqueName: \"kubernetes.io/projected/450e1ecf-5ae7-48b5-b567-e530e254f673-kube-api-access-vh2nd\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.639683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-internal-tls-certs\") pod \"placement-6fc994c476-fv9c9\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.662903 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.662941 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.678604 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.741004 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.745659 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.748810 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 18:23:18 crc kubenswrapper[4926]: I0312 18:23:18.764823 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:19 crc kubenswrapper[4926]: I0312 18:23:19.240301 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 18:23:19 crc kubenswrapper[4926]: I0312 18:23:19.240364 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:19 crc kubenswrapper[4926]: I0312 18:23:19.240376 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:19 crc kubenswrapper[4926]: I0312 18:23:19.240385 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 18:23:22 crc kubenswrapper[4926]: I0312 18:23:22.046011 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 18:23:22 crc kubenswrapper[4926]: I0312 18:23:22.243556 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:22 crc kubenswrapper[4926]: I0312 18:23:22.423172 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7pqqh"] Mar 12 18:23:22 crc kubenswrapper[4926]: I0312 18:23:22.423391 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" podUID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerName="dnsmasq-dns" containerID="cri-o://acee92af6092ee87ba774fbb0a9ef600de19a9fc8c8b93c392fa1761ce8a5481" gracePeriod=10 Mar 12 18:23:22 crc kubenswrapper[4926]: I0312 18:23:22.586217 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.058485 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.185225 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.217676 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.217727 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.295634 4926 generic.go:334] "Generic (PLEG): container finished" podID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerID="acee92af6092ee87ba774fbb0a9ef600de19a9fc8c8b93c392fa1761ce8a5481" exitCode=0 Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.296418 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" event={"ID":"af5b15ef-fa73-4ef4-9235-41da81503d2c","Type":"ContainerDied","Data":"acee92af6092ee87ba774fbb0a9ef600de19a9fc8c8b93c392fa1761ce8a5481"} Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.317917 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.319556 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.959953 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:23 crc kubenswrapper[4926]: I0312 18:23:23.971579 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.042698 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-combined-ca-bundle\") pod \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.042752 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-combined-ca-bundle\") pod \"af5704dd-cd13-4e5f-a77b-01266c63eeba\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.042806 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-db-sync-config-data\") pod \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.042849 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-scripts\") pod \"af5704dd-cd13-4e5f-a77b-01266c63eeba\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.042866 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-config-data\") pod \"af5704dd-cd13-4e5f-a77b-01266c63eeba\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.043047 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-fernet-keys\") pod \"af5704dd-cd13-4e5f-a77b-01266c63eeba\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.043074 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-credential-keys\") pod \"af5704dd-cd13-4e5f-a77b-01266c63eeba\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.043098 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbwc8\" (UniqueName: \"kubernetes.io/projected/af5704dd-cd13-4e5f-a77b-01266c63eeba-kube-api-access-zbwc8\") pod \"af5704dd-cd13-4e5f-a77b-01266c63eeba\" (UID: \"af5704dd-cd13-4e5f-a77b-01266c63eeba\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.043117 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mklss\" (UniqueName: \"kubernetes.io/projected/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-kube-api-access-mklss\") pod \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\" (UID: \"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.048206 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-scripts" (OuterVolumeSpecName: "scripts") pod "af5704dd-cd13-4e5f-a77b-01266c63eeba" (UID: "af5704dd-cd13-4e5f-a77b-01266c63eeba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.061049 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "af5704dd-cd13-4e5f-a77b-01266c63eeba" (UID: "af5704dd-cd13-4e5f-a77b-01266c63eeba"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.076598 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-kube-api-access-mklss" (OuterVolumeSpecName: "kube-api-access-mklss") pod "e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" (UID: "e96fcb3d-2f9f-468d-bafa-060a9d1f1af6"). InnerVolumeSpecName "kube-api-access-mklss". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.081860 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" (UID: "e96fcb3d-2f9f-468d-bafa-060a9d1f1af6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.100635 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "af5704dd-cd13-4e5f-a77b-01266c63eeba" (UID: "af5704dd-cd13-4e5f-a77b-01266c63eeba"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.101118 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af5704dd-cd13-4e5f-a77b-01266c63eeba-kube-api-access-zbwc8" (OuterVolumeSpecName: "kube-api-access-zbwc8") pod "af5704dd-cd13-4e5f-a77b-01266c63eeba" (UID: "af5704dd-cd13-4e5f-a77b-01266c63eeba"). InnerVolumeSpecName "kube-api-access-zbwc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.101586 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" (UID: "e96fcb3d-2f9f-468d-bafa-060a9d1f1af6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.107059 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af5704dd-cd13-4e5f-a77b-01266c63eeba" (UID: "af5704dd-cd13-4e5f-a77b-01266c63eeba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.133221 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-config-data" (OuterVolumeSpecName: "config-data") pod "af5704dd-cd13-4e5f-a77b-01266c63eeba" (UID: "af5704dd-cd13-4e5f-a77b-01266c63eeba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.144963 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145037 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145569 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145644 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145703 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145756 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145810 4926 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145861 4926 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/af5704dd-cd13-4e5f-a77b-01266c63eeba-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145911 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbwc8\" (UniqueName: \"kubernetes.io/projected/af5704dd-cd13-4e5f-a77b-01266c63eeba-kube-api-access-zbwc8\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.145969 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mklss\" (UniqueName: \"kubernetes.io/projected/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6-kube-api-access-mklss\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.246634 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-sb\") pod \"af5b15ef-fa73-4ef4-9235-41da81503d2c\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.246720 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config\") pod \"af5b15ef-fa73-4ef4-9235-41da81503d2c\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.246751 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-svc\") pod \"af5b15ef-fa73-4ef4-9235-41da81503d2c\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.246792 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-swift-storage-0\") pod \"af5b15ef-fa73-4ef4-9235-41da81503d2c\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.246832 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzlwc\" (UniqueName: \"kubernetes.io/projected/af5b15ef-fa73-4ef4-9235-41da81503d2c-kube-api-access-gzlwc\") pod \"af5b15ef-fa73-4ef4-9235-41da81503d2c\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.246867 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-nb\") pod \"af5b15ef-fa73-4ef4-9235-41da81503d2c\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.260724 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af5b15ef-fa73-4ef4-9235-41da81503d2c-kube-api-access-gzlwc" (OuterVolumeSpecName: "kube-api-access-gzlwc") pod "af5b15ef-fa73-4ef4-9235-41da81503d2c" (UID: "af5b15ef-fa73-4ef4-9235-41da81503d2c"). InnerVolumeSpecName "kube-api-access-gzlwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.292889 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "af5b15ef-fa73-4ef4-9235-41da81503d2c" (UID: "af5b15ef-fa73-4ef4-9235-41da81503d2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.307128 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "af5b15ef-fa73-4ef4-9235-41da81503d2c" (UID: "af5b15ef-fa73-4ef4-9235-41da81503d2c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.307142 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "af5b15ef-fa73-4ef4-9235-41da81503d2c" (UID: "af5b15ef-fa73-4ef4-9235-41da81503d2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.323170 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "af5b15ef-fa73-4ef4-9235-41da81503d2c" (UID: "af5b15ef-fa73-4ef4-9235-41da81503d2c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.326054 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sgnbr" event={"ID":"e96fcb3d-2f9f-468d-bafa-060a9d1f1af6","Type":"ContainerDied","Data":"89660f246e3c30bdbd43247de70865e18df164fd93b36ad1830555d68b2f5032"} Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.326093 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89660f246e3c30bdbd43247de70865e18df164fd93b36ad1830555d68b2f5032" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.326146 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sgnbr" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.333983 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerStarted","Data":"a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b"} Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.336023 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-98lfj" event={"ID":"af5704dd-cd13-4e5f-a77b-01266c63eeba","Type":"ContainerDied","Data":"45ccbdc85c6be068623b090cab069fbadeba37022e035780f13ac08ff104b1bc"} Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.336057 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45ccbdc85c6be068623b090cab069fbadeba37022e035780f13ac08ff104b1bc" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.336115 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-98lfj" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348399 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config" (OuterVolumeSpecName: "config") pod "af5b15ef-fa73-4ef4-9235-41da81503d2c" (UID: "af5b15ef-fa73-4ef4-9235-41da81503d2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348526 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config\") pod \"af5b15ef-fa73-4ef4-9235-41da81503d2c\" (UID: \"af5b15ef-fa73-4ef4-9235-41da81503d2c\") " Mar 12 18:23:24 crc kubenswrapper[4926]: W0312 18:23:24.348657 4926 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/af5b15ef-fa73-4ef4-9235-41da81503d2c/volumes/kubernetes.io~configmap/config Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348672 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config" (OuterVolumeSpecName: "config") pod "af5b15ef-fa73-4ef4-9235-41da81503d2c" (UID: "af5b15ef-fa73-4ef4-9235-41da81503d2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348907 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348921 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-7pqqh" event={"ID":"af5b15ef-fa73-4ef4-9235-41da81503d2c","Type":"ContainerDied","Data":"c327e0d0fbf885a0bb6f202d846326d2682c20d15307d1ed071e6aa3068c47c0"} Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348965 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348981 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348991 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.348999 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.349002 4926 scope.go:117] "RemoveContainer" containerID="acee92af6092ee87ba774fbb0a9ef600de19a9fc8c8b93c392fa1761ce8a5481" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.349009 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af5b15ef-fa73-4ef4-9235-41da81503d2c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.349020 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzlwc\" (UniqueName: \"kubernetes.io/projected/af5b15ef-fa73-4ef4-9235-41da81503d2c-kube-api-access-gzlwc\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.404998 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7pqqh"] Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.407314 4926 scope.go:117] "RemoveContainer" containerID="d817244fbaa6cac20d7a51e9c98e06316cbeae770e3def99741dd49bb46b1970" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.419557 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-7pqqh"] Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.500357 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af5b15ef-fa73-4ef4-9235-41da81503d2c" path="/var/lib/kubelet/pods/af5b15ef-fa73-4ef4-9235-41da81503d2c/volumes" Mar 12 18:23:24 crc kubenswrapper[4926]: I0312 18:23:24.515282 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fc994c476-fv9c9"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.085681 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5b667c464b-fk8sc"] Mar 12 18:23:25 crc kubenswrapper[4926]: E0312 18:23:25.086971 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" containerName="barbican-db-sync" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.087015 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" containerName="barbican-db-sync" Mar 12 18:23:25 crc kubenswrapper[4926]: E0312 18:23:25.087051 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerName="dnsmasq-dns" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.087061 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerName="dnsmasq-dns" Mar 12 18:23:25 crc kubenswrapper[4926]: E0312 18:23:25.087103 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerName="init" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.087111 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerName="init" Mar 12 18:23:25 crc kubenswrapper[4926]: E0312 18:23:25.087125 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af5704dd-cd13-4e5f-a77b-01266c63eeba" containerName="keystone-bootstrap" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.087133 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af5704dd-cd13-4e5f-a77b-01266c63eeba" containerName="keystone-bootstrap" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.087357 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="af5b15ef-fa73-4ef4-9235-41da81503d2c" containerName="dnsmasq-dns" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.087384 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="af5704dd-cd13-4e5f-a77b-01266c63eeba" containerName="keystone-bootstrap" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.087401 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" containerName="barbican-db-sync" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.088183 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.089927 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.092234 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.092237 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.092759 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-vrsc7" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.093013 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.093098 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.105401 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b667c464b-fk8sc"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.189403 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-scripts\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.189488 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-combined-ca-bundle\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.189509 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-internal-tls-certs\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.189544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-fernet-keys\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.189634 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-public-tls-certs\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.189793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdgrq\" (UniqueName: \"kubernetes.io/projected/054d27f0-5c9b-4e59-98b3-e05609c3b257-kube-api-access-bdgrq\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.189985 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-credential-keys\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.190032 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-config-data\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.233725 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-78dfff8dc9-fgs8s"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.235078 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.238204 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.238370 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hz28b" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.238600 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.293487 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-credential-keys\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.293723 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-config-data\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.293834 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-scripts\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.294037 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-combined-ca-bundle\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.294104 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-internal-tls-certs\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.294201 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-fernet-keys\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.294264 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-public-tls-certs\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.294364 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdgrq\" (UniqueName: \"kubernetes.io/projected/054d27f0-5c9b-4e59-98b3-e05609c3b257-kube-api-access-bdgrq\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.306098 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-config-data\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.308914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-public-tls-certs\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.309331 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-combined-ca-bundle\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.309353 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-credential-keys\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.312048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-fernet-keys\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.312849 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-internal-tls-certs\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.318878 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/054d27f0-5c9b-4e59-98b3-e05609c3b257-scripts\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.352812 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78dfff8dc9-fgs8s"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.354197 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdgrq\" (UniqueName: \"kubernetes.io/projected/054d27f0-5c9b-4e59-98b3-e05609c3b257-kube-api-access-bdgrq\") pod \"keystone-5b667c464b-fk8sc\" (UID: \"054d27f0-5c9b-4e59-98b3-e05609c3b257\") " pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.396404 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.396472 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6qc2\" (UniqueName: \"kubernetes.io/projected/81afb8cd-af2c-4515-b4d1-893903371af0-kube-api-access-v6qc2\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.396530 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-combined-ca-bundle\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.396583 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81afb8cd-af2c-4515-b4d1-893903371af0-logs\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.396660 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data-custom\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.410345 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fc994c476-fv9c9" event={"ID":"450e1ecf-5ae7-48b5-b567-e530e254f673","Type":"ContainerStarted","Data":"e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73"} Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.410387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fc994c476-fv9c9" event={"ID":"450e1ecf-5ae7-48b5-b567-e530e254f673","Type":"ContainerStarted","Data":"cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a"} Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.410396 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fc994c476-fv9c9" event={"ID":"450e1ecf-5ae7-48b5-b567-e530e254f673","Type":"ContainerStarted","Data":"1f1ee7c53345b2a1183a515645bed6bfd508f8efad3f3b4a2588d4c7787c2e09"} Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.411398 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.411673 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.422750 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.456247 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-94956785d-mtl2w"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.475874 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.482369 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.491916 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-gbmf8"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.500590 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.500626 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6qc2\" (UniqueName: \"kubernetes.io/projected/81afb8cd-af2c-4515-b4d1-893903371af0-kube-api-access-v6qc2\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.500687 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-combined-ca-bundle\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.500755 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81afb8cd-af2c-4515-b4d1-893903371af0-logs\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.500794 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data-custom\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.502730 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.504056 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81afb8cd-af2c-4515-b4d1-893903371af0-logs\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.514863 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.515340 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data-custom\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.515676 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-fddddf9f9-kbftb"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.526664 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-combined-ca-bundle\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.528597 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.537734 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6qc2\" (UniqueName: \"kubernetes.io/projected/81afb8cd-af2c-4515-b4d1-893903371af0-kube-api-access-v6qc2\") pod \"barbican-worker-78dfff8dc9-fgs8s\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.540630 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-94956785d-mtl2w"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.556904 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.595033 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fddddf9f9-kbftb"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.604355 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.604421 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ktzb\" (UniqueName: \"kubernetes.io/projected/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-kube-api-access-8ktzb\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.605928 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-config-data\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.605967 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-config-data-custom\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.605989 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-svc\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606021 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da936793-13b1-4815-a1ec-4d5d609ca5e3-logs\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606038 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m7fh\" (UniqueName: \"kubernetes.io/projected/da936793-13b1-4815-a1ec-4d5d609ca5e3-kube-api-access-2m7fh\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606080 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxtxm\" (UniqueName: \"kubernetes.io/projected/08abfe56-0e5c-4634-9a1a-488e2bbb587d-kube-api-access-pxtxm\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606099 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-combined-ca-bundle\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606122 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606145 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-config\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606180 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-combined-ca-bundle\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606199 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data-custom\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606243 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606262 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.606307 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-logs\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.640537 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-gbmf8"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.679439 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6fc994c476-fv9c9" podStartSLOduration=7.679414865 podStartE2EDuration="7.679414865s" podCreationTimestamp="2026-03-12 18:23:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:25.640875686 +0000 UTC m=+1246.009502019" watchObservedRunningTime="2026-03-12 18:23:25.679414865 +0000 UTC m=+1246.048041198" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.680504 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-74668b896d-mctls"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.681986 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709498 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ktzb\" (UniqueName: \"kubernetes.io/projected/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-kube-api-access-8ktzb\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709533 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-config-data\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709553 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-config-data-custom\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709574 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-svc\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709601 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da936793-13b1-4815-a1ec-4d5d609ca5e3-logs\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709616 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m7fh\" (UniqueName: \"kubernetes.io/projected/da936793-13b1-4815-a1ec-4d5d609ca5e3-kube-api-access-2m7fh\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709640 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxtxm\" (UniqueName: \"kubernetes.io/projected/08abfe56-0e5c-4634-9a1a-488e2bbb587d-kube-api-access-pxtxm\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-combined-ca-bundle\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709677 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709700 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-config\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709728 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-combined-ca-bundle\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709749 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data-custom\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709776 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709794 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709830 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-logs\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.709855 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.710552 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.713090 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-config\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.713392 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da936793-13b1-4815-a1ec-4d5d609ca5e3-logs\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.717144 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.717910 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.723874 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-logs\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.734623 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-svc\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.739277 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-config-data\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.741494 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.741558 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-74668b896d-mctls"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.742853 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data-custom\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.743188 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-combined-ca-bundle\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.743321 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-config-data-custom\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.750661 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxtxm\" (UniqueName: \"kubernetes.io/projected/08abfe56-0e5c-4634-9a1a-488e2bbb587d-kube-api-access-pxtxm\") pod \"dnsmasq-dns-85ff748b95-gbmf8\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.756469 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-combined-ca-bundle\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.762129 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ktzb\" (UniqueName: \"kubernetes.io/projected/08cef0ec-16bc-4b64-95f6-e0d8f22fa00e-kube-api-access-8ktzb\") pod \"barbican-worker-fddddf9f9-kbftb\" (UID: \"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e\") " pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.778598 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m7fh\" (UniqueName: \"kubernetes.io/projected/da936793-13b1-4815-a1ec-4d5d609ca5e3-kube-api-access-2m7fh\") pod \"barbican-keystone-listener-94956785d-mtl2w\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.811999 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-config-data-custom\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.812090 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-logs\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.812110 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfzr5\" (UniqueName: \"kubernetes.io/projected/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-kube-api-access-wfzr5\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.812131 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-combined-ca-bundle\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.812253 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-config-data\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.831852 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.872966 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9b55c586b-s7wqs"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.874458 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.892280 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.892810 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.908179 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-fddddf9f9-kbftb" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.914431 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-logs\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.914498 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfzr5\" (UniqueName: \"kubernetes.io/projected/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-kube-api-access-wfzr5\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.914524 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-combined-ca-bundle\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.914744 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-config-data\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.914843 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-config-data-custom\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.915351 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-logs\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.916932 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9b55c586b-s7wqs"] Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.931567 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-config-data\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.954413 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-combined-ca-bundle\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.982737 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfzr5\" (UniqueName: \"kubernetes.io/projected/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-kube-api-access-wfzr5\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:25 crc kubenswrapper[4926]: I0312 18:23:25.984957 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92-config-data-custom\") pod \"barbican-keystone-listener-74668b896d-mctls\" (UID: \"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92\") " pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.016748 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qldmj\" (UniqueName: \"kubernetes.io/projected/2f1f2a42-878e-46c0-bd66-4927a4689299-kube-api-access-qldmj\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.016831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1f2a42-878e-46c0-bd66-4927a4689299-logs\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.016881 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.017038 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-combined-ca-bundle\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.017064 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data-custom\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.019068 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-74668b896d-mctls" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.122573 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-combined-ca-bundle\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.122651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data-custom\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.122688 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qldmj\" (UniqueName: \"kubernetes.io/projected/2f1f2a42-878e-46c0-bd66-4927a4689299-kube-api-access-qldmj\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.122728 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1f2a42-878e-46c0-bd66-4927a4689299-logs\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.122764 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.126995 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1f2a42-878e-46c0-bd66-4927a4689299-logs\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.138228 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.145968 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-combined-ca-bundle\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.149071 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qldmj\" (UniqueName: \"kubernetes.io/projected/2f1f2a42-878e-46c0-bd66-4927a4689299-kube-api-access-qldmj\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.149553 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data-custom\") pod \"barbican-api-9b55c586b-s7wqs\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.224456 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.475927 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lgvzs" event={"ID":"dac4b5d6-fb31-4955-8679-db9d3ff63c10","Type":"ContainerStarted","Data":"96adbf85f267b1793c500e8c152eb5c14d98f7080edf6f6f116872d3c431cb0a"} Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.501929 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-lgvzs" podStartSLOduration=3.318637497 podStartE2EDuration="42.501911225s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="2026-03-12 18:22:45.744242434 +0000 UTC m=+1206.112868767" lastFinishedPulling="2026-03-12 18:23:24.927516162 +0000 UTC m=+1245.296142495" observedRunningTime="2026-03-12 18:23:26.496739203 +0000 UTC m=+1246.865365536" watchObservedRunningTime="2026-03-12 18:23:26.501911225 +0000 UTC m=+1246.870537558" Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.582195 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5b667c464b-fk8sc"] Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.600199 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78dfff8dc9-fgs8s"] Mar 12 18:23:26 crc kubenswrapper[4926]: W0312 18:23:26.621093 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod054d27f0_5c9b_4e59_98b3_e05609c3b257.slice/crio-f3c3a12240b2c31f1ffa837e04ed066879a72f603a656ed2b36d6c9e4a650f2f WatchSource:0}: Error finding container f3c3a12240b2c31f1ffa837e04ed066879a72f603a656ed2b36d6c9e4a650f2f: Status 404 returned error can't find the container with id f3c3a12240b2c31f1ffa837e04ed066879a72f603a656ed2b36d6c9e4a650f2f Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.736911 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-94956785d-mtl2w"] Mar 12 18:23:26 crc kubenswrapper[4926]: W0312 18:23:26.744761 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda936793_13b1_4815_a1ec_4d5d609ca5e3.slice/crio-01f48ef206a2d7f242f74c69621856e49171cd0d33097cfbec64d9c1935505bd WatchSource:0}: Error finding container 01f48ef206a2d7f242f74c69621856e49171cd0d33097cfbec64d9c1935505bd: Status 404 returned error can't find the container with id 01f48ef206a2d7f242f74c69621856e49171cd0d33097cfbec64d9c1935505bd Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.921781 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-gbmf8"] Mar 12 18:23:26 crc kubenswrapper[4926]: I0312 18:23:26.935134 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-74668b896d-mctls"] Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.099260 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9b55c586b-s7wqs"] Mar 12 18:23:27 crc kubenswrapper[4926]: W0312 18:23:27.118364 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f1f2a42_878e_46c0_bd66_4927a4689299.slice/crio-28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc WatchSource:0}: Error finding container 28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc: Status 404 returned error can't find the container with id 28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.123723 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-fddddf9f9-kbftb"] Mar 12 18:23:27 crc kubenswrapper[4926]: W0312 18:23:27.130096 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08cef0ec_16bc_4b64_95f6_e0d8f22fa00e.slice/crio-c8e2f3a7bbe0bf79a104ee85f894f11f7f0fba578b28b0c57b6258b01319863e WatchSource:0}: Error finding container c8e2f3a7bbe0bf79a104ee85f894f11f7f0fba578b28b0c57b6258b01319863e: Status 404 returned error can't find the container with id c8e2f3a7bbe0bf79a104ee85f894f11f7f0fba578b28b0c57b6258b01319863e Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.487403 4926 generic.go:334] "Generic (PLEG): container finished" podID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerID="293a2572fcb8b2d40b187537e1c594705c232add90d5f330faf45d6e18e49dd0" exitCode=0 Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.487493 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" event={"ID":"08abfe56-0e5c-4634-9a1a-488e2bbb587d","Type":"ContainerDied","Data":"293a2572fcb8b2d40b187537e1c594705c232add90d5f330faf45d6e18e49dd0"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.487912 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" event={"ID":"08abfe56-0e5c-4634-9a1a-488e2bbb587d","Type":"ContainerStarted","Data":"07e4278fe40ae8695fb94d6a2d5bec78c86b69a3255ccbd19aa22a7865e54876"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.493962 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" event={"ID":"81afb8cd-af2c-4515-b4d1-893903371af0","Type":"ContainerStarted","Data":"3adbb909a2691a9c3475b5b8bcf70c23ddd0ca36e2352ef1830b72181f72ded0"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.496361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b667c464b-fk8sc" event={"ID":"054d27f0-5c9b-4e59-98b3-e05609c3b257","Type":"ContainerStarted","Data":"272c18702618eb570c10d215f82e06e4aaeb5cf9a9534f622edc7dd3c1a2f82b"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.496398 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5b667c464b-fk8sc" event={"ID":"054d27f0-5c9b-4e59-98b3-e05609c3b257","Type":"ContainerStarted","Data":"f3c3a12240b2c31f1ffa837e04ed066879a72f603a656ed2b36d6c9e4a650f2f"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.497410 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.509276 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74668b896d-mctls" event={"ID":"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92","Type":"ContainerStarted","Data":"7ab0fc4ee2052252ea1c5a664f1849480ecb5c7e93f1eac70fd366eabd6157fb"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.511048 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fddddf9f9-kbftb" event={"ID":"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e","Type":"ContainerStarted","Data":"c8e2f3a7bbe0bf79a104ee85f894f11f7f0fba578b28b0c57b6258b01319863e"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.514275 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" event={"ID":"da936793-13b1-4815-a1ec-4d5d609ca5e3","Type":"ContainerStarted","Data":"01f48ef206a2d7f242f74c69621856e49171cd0d33097cfbec64d9c1935505bd"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.519157 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9b55c586b-s7wqs" event={"ID":"2f1f2a42-878e-46c0-bd66-4927a4689299","Type":"ContainerStarted","Data":"0ae594d417f33a2af41173b2f1260c8120ad7232590b0279e7d83cb0dbe67a9c"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.519198 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9b55c586b-s7wqs" event={"ID":"2f1f2a42-878e-46c0-bd66-4927a4689299","Type":"ContainerStarted","Data":"28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc"} Mar 12 18:23:27 crc kubenswrapper[4926]: I0312 18:23:27.532790 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5b667c464b-fk8sc" podStartSLOduration=2.532775262 podStartE2EDuration="2.532775262s" podCreationTimestamp="2026-03-12 18:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:27.5260103 +0000 UTC m=+1247.894636663" watchObservedRunningTime="2026-03-12 18:23:27.532775262 +0000 UTC m=+1247.901401595" Mar 12 18:23:28 crc kubenswrapper[4926]: I0312 18:23:28.529513 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" event={"ID":"08abfe56-0e5c-4634-9a1a-488e2bbb587d","Type":"ContainerStarted","Data":"3e365f4cb50835b1180ff7c19d7fc397df5fff675446efdb00f5237496ddec2b"} Mar 12 18:23:28 crc kubenswrapper[4926]: I0312 18:23:28.530982 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:28 crc kubenswrapper[4926]: I0312 18:23:28.533739 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9b55c586b-s7wqs" event={"ID":"2f1f2a42-878e-46c0-bd66-4927a4689299","Type":"ContainerStarted","Data":"8231f1ef7dfab85032bd8e8cc8a2434c72aeff5f8977a2307df4d52b313ce335"} Mar 12 18:23:28 crc kubenswrapper[4926]: I0312 18:23:28.533777 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:28 crc kubenswrapper[4926]: I0312 18:23:28.533801 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:28 crc kubenswrapper[4926]: I0312 18:23:28.558189 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" podStartSLOduration=3.558164808 podStartE2EDuration="3.558164808s" podCreationTimestamp="2026-03-12 18:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:28.552063996 +0000 UTC m=+1248.920690349" watchObservedRunningTime="2026-03-12 18:23:28.558164808 +0000 UTC m=+1248.926791141" Mar 12 18:23:28 crc kubenswrapper[4926]: I0312 18:23:28.586692 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9b55c586b-s7wqs" podStartSLOduration=3.586674232 podStartE2EDuration="3.586674232s" podCreationTimestamp="2026-03-12 18:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:28.576955117 +0000 UTC m=+1248.945581460" watchObservedRunningTime="2026-03-12 18:23:28.586674232 +0000 UTC m=+1248.955300565" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.297702 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6fd6447dc6-7dg85"] Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.299094 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.304604 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.305568 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.332607 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fd6447dc6-7dg85"] Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.402791 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-config-data\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.403346 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-combined-ca-bundle\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.403541 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-internal-tls-certs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.403624 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-public-tls-certs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.403685 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-config-data-custom\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.403745 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc7tt\" (UniqueName: \"kubernetes.io/projected/32458d94-2727-4718-a842-f20e68b6a0dd-kube-api-access-kc7tt\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.403821 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32458d94-2727-4718-a842-f20e68b6a0dd-logs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.505751 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-public-tls-certs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.505837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-config-data-custom\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.505887 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc7tt\" (UniqueName: \"kubernetes.io/projected/32458d94-2727-4718-a842-f20e68b6a0dd-kube-api-access-kc7tt\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.505933 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32458d94-2727-4718-a842-f20e68b6a0dd-logs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.505988 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-config-data\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.506031 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-combined-ca-bundle\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.506085 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-internal-tls-certs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.507035 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/32458d94-2727-4718-a842-f20e68b6a0dd-logs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.514247 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-combined-ca-bundle\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.529941 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-public-tls-certs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.531494 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-internal-tls-certs\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.533295 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-config-data\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.535722 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/32458d94-2727-4718-a842-f20e68b6a0dd-config-data-custom\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.544482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc7tt\" (UniqueName: \"kubernetes.io/projected/32458d94-2727-4718-a842-f20e68b6a0dd-kube-api-access-kc7tt\") pod \"barbican-api-6fd6447dc6-7dg85\" (UID: \"32458d94-2727-4718-a842-f20e68b6a0dd\") " pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:29 crc kubenswrapper[4926]: I0312 18:23:29.631663 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.450421 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6fd6447dc6-7dg85"] Mar 12 18:23:30 crc kubenswrapper[4926]: W0312 18:23:30.466797 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32458d94_2727_4718_a842_f20e68b6a0dd.slice/crio-9c7e7e89198f2cf9ac44a3023e839fc1bfe908ca6727788ccbb5e057b4856f40 WatchSource:0}: Error finding container 9c7e7e89198f2cf9ac44a3023e839fc1bfe908ca6727788ccbb5e057b4856f40: Status 404 returned error can't find the container with id 9c7e7e89198f2cf9ac44a3023e839fc1bfe908ca6727788ccbb5e057b4856f40 Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.568647 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" event={"ID":"81afb8cd-af2c-4515-b4d1-893903371af0","Type":"ContainerStarted","Data":"300d112166ede447ec6708957fbc43526b8d0fa27de369019ba8b434782269d0"} Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.568700 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" event={"ID":"81afb8cd-af2c-4515-b4d1-893903371af0","Type":"ContainerStarted","Data":"bee09196dcaf75cb5eadfa250f0ca7febdf9ab076b98f039b2a754122b989aef"} Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.582147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74668b896d-mctls" event={"ID":"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92","Type":"ContainerStarted","Data":"0c2e139904b5d1ac08e504463fb8ce4735df2887bdf04c680ede7cd5ab04f211"} Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.585194 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fddddf9f9-kbftb" event={"ID":"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e","Type":"ContainerStarted","Data":"d5a36d4293aef7b5d2a9d41d34927260eb58046868118cf247fcac4acce3f566"} Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.586196 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fd6447dc6-7dg85" event={"ID":"32458d94-2727-4718-a842-f20e68b6a0dd","Type":"ContainerStarted","Data":"9c7e7e89198f2cf9ac44a3023e839fc1bfe908ca6727788ccbb5e057b4856f40"} Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.590069 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" event={"ID":"da936793-13b1-4815-a1ec-4d5d609ca5e3","Type":"ContainerStarted","Data":"38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2"} Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.631315 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-fddddf9f9-kbftb" podStartSLOduration=2.816225767 podStartE2EDuration="5.63129129s" podCreationTimestamp="2026-03-12 18:23:25 +0000 UTC" firstStartedPulling="2026-03-12 18:23:27.140723961 +0000 UTC m=+1247.509350294" lastFinishedPulling="2026-03-12 18:23:29.955789484 +0000 UTC m=+1250.324415817" observedRunningTime="2026-03-12 18:23:30.611283002 +0000 UTC m=+1250.979909335" watchObservedRunningTime="2026-03-12 18:23:30.63129129 +0000 UTC m=+1250.999917623" Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.678167 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" podStartSLOduration=2.38227452 podStartE2EDuration="5.67814163s" podCreationTimestamp="2026-03-12 18:23:25 +0000 UTC" firstStartedPulling="2026-03-12 18:23:26.661298566 +0000 UTC m=+1247.029924899" lastFinishedPulling="2026-03-12 18:23:29.957165676 +0000 UTC m=+1250.325792009" observedRunningTime="2026-03-12 18:23:30.637233397 +0000 UTC m=+1251.005859730" watchObservedRunningTime="2026-03-12 18:23:30.67814163 +0000 UTC m=+1251.046767963" Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.737066 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-78dfff8dc9-fgs8s"] Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.742710 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-74668b896d-mctls" podStartSLOduration=2.72742515 podStartE2EDuration="5.742692056s" podCreationTimestamp="2026-03-12 18:23:25 +0000 UTC" firstStartedPulling="2026-03-12 18:23:26.938574827 +0000 UTC m=+1247.307201160" lastFinishedPulling="2026-03-12 18:23:29.953841733 +0000 UTC m=+1250.322468066" observedRunningTime="2026-03-12 18:23:30.677206981 +0000 UTC m=+1251.045833314" watchObservedRunningTime="2026-03-12 18:23:30.742692056 +0000 UTC m=+1251.111318389" Mar 12 18:23:30 crc kubenswrapper[4926]: I0312 18:23:30.756279 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-94956785d-mtl2w"] Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.604614 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" event={"ID":"da936793-13b1-4815-a1ec-4d5d609ca5e3","Type":"ContainerStarted","Data":"7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8"} Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.604696 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener-log" containerID="cri-o://38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2" gracePeriod=30 Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.604736 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener" containerID="cri-o://7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8" gracePeriod=30 Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.608998 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74668b896d-mctls" event={"ID":"4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92","Type":"ContainerStarted","Data":"c8ec8c9afa0fc5bd02807c600c4d8dde24f99e3b5f25ba0d76685c20d0c9ee08"} Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.616268 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-fddddf9f9-kbftb" event={"ID":"08cef0ec-16bc-4b64-95f6-e0d8f22fa00e","Type":"ContainerStarted","Data":"6993d35552e1e9c5edf87ca66043b892ec364d28d08d5907d5afa076e67ba45a"} Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.618480 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fd6447dc6-7dg85" event={"ID":"32458d94-2727-4718-a842-f20e68b6a0dd","Type":"ContainerStarted","Data":"ee2e4bc6fdc6d3d511e5d9d18c4d7f78ce09e67e478b5dd24044bb375fb67600"} Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.618523 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6fd6447dc6-7dg85" event={"ID":"32458d94-2727-4718-a842-f20e68b6a0dd","Type":"ContainerStarted","Data":"cedc169daa886f2a21e9492e6b54650af1b8208615b2c2bfa967597a97d9c542"} Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.646222 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" podStartSLOduration=3.441357733 podStartE2EDuration="6.646200567s" podCreationTimestamp="2026-03-12 18:23:25 +0000 UTC" firstStartedPulling="2026-03-12 18:23:26.751125005 +0000 UTC m=+1247.119751338" lastFinishedPulling="2026-03-12 18:23:29.955967839 +0000 UTC m=+1250.324594172" observedRunningTime="2026-03-12 18:23:31.621890404 +0000 UTC m=+1251.990516737" watchObservedRunningTime="2026-03-12 18:23:31.646200567 +0000 UTC m=+1252.014826900" Mar 12 18:23:31 crc kubenswrapper[4926]: I0312 18:23:31.650600 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6fd6447dc6-7dg85" podStartSLOduration=2.650582665 podStartE2EDuration="2.650582665s" podCreationTimestamp="2026-03-12 18:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:31.645486204 +0000 UTC m=+1252.014112537" watchObservedRunningTime="2026-03-12 18:23:31.650582665 +0000 UTC m=+1252.019208988" Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.636521 4926 generic.go:334] "Generic (PLEG): container finished" podID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerID="38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2" exitCode=143 Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.636572 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" event={"ID":"da936793-13b1-4815-a1ec-4d5d609ca5e3","Type":"ContainerDied","Data":"38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2"} Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.638389 4926 generic.go:334] "Generic (PLEG): container finished" podID="dac4b5d6-fb31-4955-8679-db9d3ff63c10" containerID="96adbf85f267b1793c500e8c152eb5c14d98f7080edf6f6f116872d3c431cb0a" exitCode=0 Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.638568 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lgvzs" event={"ID":"dac4b5d6-fb31-4955-8679-db9d3ff63c10","Type":"ContainerDied","Data":"96adbf85f267b1793c500e8c152eb5c14d98f7080edf6f6f116872d3c431cb0a"} Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.638978 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.639016 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.639416 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker-log" containerID="cri-o://bee09196dcaf75cb5eadfa250f0ca7febdf9ab076b98f039b2a754122b989aef" gracePeriod=30 Mar 12 18:23:32 crc kubenswrapper[4926]: I0312 18:23:32.639523 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker" containerID="cri-o://300d112166ede447ec6708957fbc43526b8d0fa27de369019ba8b434782269d0" gracePeriod=30 Mar 12 18:23:33 crc kubenswrapper[4926]: I0312 18:23:33.320096 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5c6848d8cd-cq57n" podUID="a1ae8f23-3518-430a-bbcf-e7be0cb8282e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Mar 12 18:23:33 crc kubenswrapper[4926]: I0312 18:23:33.633813 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 12 18:23:33 crc kubenswrapper[4926]: I0312 18:23:33.656871 4926 generic.go:334] "Generic (PLEG): container finished" podID="81afb8cd-af2c-4515-b4d1-893903371af0" containerID="300d112166ede447ec6708957fbc43526b8d0fa27de369019ba8b434782269d0" exitCode=0 Mar 12 18:23:33 crc kubenswrapper[4926]: I0312 18:23:33.656911 4926 generic.go:334] "Generic (PLEG): container finished" podID="81afb8cd-af2c-4515-b4d1-893903371af0" containerID="bee09196dcaf75cb5eadfa250f0ca7febdf9ab076b98f039b2a754122b989aef" exitCode=143 Mar 12 18:23:33 crc kubenswrapper[4926]: I0312 18:23:33.657052 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" event={"ID":"81afb8cd-af2c-4515-b4d1-893903371af0","Type":"ContainerDied","Data":"300d112166ede447ec6708957fbc43526b8d0fa27de369019ba8b434782269d0"} Mar 12 18:23:33 crc kubenswrapper[4926]: I0312 18:23:33.657185 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" event={"ID":"81afb8cd-af2c-4515-b4d1-893903371af0","Type":"ContainerDied","Data":"bee09196dcaf75cb5eadfa250f0ca7febdf9ab076b98f039b2a754122b989aef"} Mar 12 18:23:35 crc kubenswrapper[4926]: I0312 18:23:35.894676 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:35 crc kubenswrapper[4926]: I0312 18:23:35.977098 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rvkqm"] Mar 12 18:23:35 crc kubenswrapper[4926]: I0312 18:23:35.977348 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerName="dnsmasq-dns" containerID="cri-o://3d75e203f5f24817d9aa7e9a6cf9f743a4b6a96bd884d103fc15a23a410323b0" gracePeriod=10 Mar 12 18:23:36 crc kubenswrapper[4926]: I0312 18:23:36.688754 4926 generic.go:334] "Generic (PLEG): container finished" podID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerID="3d75e203f5f24817d9aa7e9a6cf9f743a4b6a96bd884d103fc15a23a410323b0" exitCode=0 Mar 12 18:23:36 crc kubenswrapper[4926]: I0312 18:23:36.688875 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" event={"ID":"b5cd3f65-2af4-4a38-ab87-c266452f8c5a","Type":"ContainerDied","Data":"3d75e203f5f24817d9aa7e9a6cf9f743a4b6a96bd884d103fc15a23a410323b0"} Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.240041 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: connect: connection refused" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.699480 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-lgvzs" event={"ID":"dac4b5d6-fb31-4955-8679-db9d3ff63c10","Type":"ContainerDied","Data":"1b929dd828244692447588d25c4d9ac5aab4e983cee4fcf89d743277f69d1b5d"} Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.699727 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b929dd828244692447588d25c4d9ac5aab4e983cee4fcf89d743277f69d1b5d" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.716799 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.775039 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.863862 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.909949 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-combined-ca-bundle\") pod \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.910060 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-config-data\") pod \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.910100 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-scripts\") pod \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.910139 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgqfm\" (UniqueName: \"kubernetes.io/projected/dac4b5d6-fb31-4955-8679-db9d3ff63c10-kube-api-access-fgqfm\") pod \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.910249 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dac4b5d6-fb31-4955-8679-db9d3ff63c10-etc-machine-id\") pod \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.910347 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-db-sync-config-data\") pod \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\" (UID: \"dac4b5d6-fb31-4955-8679-db9d3ff63c10\") " Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.913768 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dac4b5d6-fb31-4955-8679-db9d3ff63c10-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dac4b5d6-fb31-4955-8679-db9d3ff63c10" (UID: "dac4b5d6-fb31-4955-8679-db9d3ff63c10"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.933104 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dac4b5d6-fb31-4955-8679-db9d3ff63c10-kube-api-access-fgqfm" (OuterVolumeSpecName: "kube-api-access-fgqfm") pod "dac4b5d6-fb31-4955-8679-db9d3ff63c10" (UID: "dac4b5d6-fb31-4955-8679-db9d3ff63c10"). InnerVolumeSpecName "kube-api-access-fgqfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.933472 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-scripts" (OuterVolumeSpecName: "scripts") pod "dac4b5d6-fb31-4955-8679-db9d3ff63c10" (UID: "dac4b5d6-fb31-4955-8679-db9d3ff63c10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.933526 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "dac4b5d6-fb31-4955-8679-db9d3ff63c10" (UID: "dac4b5d6-fb31-4955-8679-db9d3ff63c10"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:37 crc kubenswrapper[4926]: I0312 18:23:37.960042 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dac4b5d6-fb31-4955-8679-db9d3ff63c10" (UID: "dac4b5d6-fb31-4955-8679-db9d3ff63c10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:37.999208 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-config-data" (OuterVolumeSpecName: "config-data") pod "dac4b5d6-fb31-4955-8679-db9d3ff63c10" (UID: "dac4b5d6-fb31-4955-8679-db9d3ff63c10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:38.013207 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:38.013240 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:38.013249 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:38.013260 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgqfm\" (UniqueName: \"kubernetes.io/projected/dac4b5d6-fb31-4955-8679-db9d3ff63c10-kube-api-access-fgqfm\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:38.013270 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dac4b5d6-fb31-4955-8679-db9d3ff63c10-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:38.013279 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/dac4b5d6-fb31-4955-8679-db9d3ff63c10-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:38 crc kubenswrapper[4926]: E0312 18:23:38.612509 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddac4b5d6_fb31_4955_8679_db9d3ff63c10.slice/crio-1b929dd828244692447588d25c4d9ac5aab4e983cee4fcf89d743277f69d1b5d\": RecentStats: unable to find data in memory cache]" Mar 12 18:23:38 crc kubenswrapper[4926]: I0312 18:23:38.705908 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-lgvzs" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.091687 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:39 crc kubenswrapper[4926]: E0312 18:23:39.092614 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dac4b5d6-fb31-4955-8679-db9d3ff63c10" containerName="cinder-db-sync" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.092627 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dac4b5d6-fb31-4955-8679-db9d3ff63c10" containerName="cinder-db-sync" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.092828 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dac4b5d6-fb31-4955-8679-db9d3ff63c10" containerName="cinder-db-sync" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.093690 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.098820 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.099086 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-m6xv4" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.099202 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.099703 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.165710 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.168929 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.182854 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rln7t"] Mar 12 18:23:39 crc kubenswrapper[4926]: E0312 18:23:39.183259 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker-log" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.183271 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker-log" Mar 12 18:23:39 crc kubenswrapper[4926]: E0312 18:23:39.183313 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.183321 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.183495 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.183510 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" containerName="barbican-worker-log" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.184394 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.209754 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rln7t"] Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.214365 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.241429 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6qc2\" (UniqueName: \"kubernetes.io/projected/81afb8cd-af2c-4515-b4d1-893903371af0-kube-api-access-v6qc2\") pod \"81afb8cd-af2c-4515-b4d1-893903371af0\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.241592 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81afb8cd-af2c-4515-b4d1-893903371af0-logs\") pod \"81afb8cd-af2c-4515-b4d1-893903371af0\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.241643 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data-custom\") pod \"81afb8cd-af2c-4515-b4d1-893903371af0\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.241768 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-combined-ca-bundle\") pod \"81afb8cd-af2c-4515-b4d1-893903371af0\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.241806 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data\") pod \"81afb8cd-af2c-4515-b4d1-893903371af0\" (UID: \"81afb8cd-af2c-4515-b4d1-893903371af0\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242103 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242128 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242198 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242222 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242251 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpn6w\" (UniqueName: \"kubernetes.io/projected/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-kube-api-access-jpn6w\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242279 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242344 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm8xc\" (UniqueName: \"kubernetes.io/projected/75ab53a0-e811-4b31-8a1e-d71be4115b2c-kube-api-access-sm8xc\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242388 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242421 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-config\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242500 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242582 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.242605 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.243991 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81afb8cd-af2c-4515-b4d1-893903371af0-logs" (OuterVolumeSpecName: "logs") pod "81afb8cd-af2c-4515-b4d1-893903371af0" (UID: "81afb8cd-af2c-4515-b4d1-893903371af0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.261583 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "81afb8cd-af2c-4515-b4d1-893903371af0" (UID: "81afb8cd-af2c-4515-b4d1-893903371af0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.263845 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81afb8cd-af2c-4515-b4d1-893903371af0-kube-api-access-v6qc2" (OuterVolumeSpecName: "kube-api-access-v6qc2") pod "81afb8cd-af2c-4515-b4d1-893903371af0" (UID: "81afb8cd-af2c-4515-b4d1-893903371af0"). InnerVolumeSpecName "kube-api-access-v6qc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.318144 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:39 crc kubenswrapper[4926]: E0312 18:23:39.318562 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerName="dnsmasq-dns" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.318594 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerName="dnsmasq-dns" Mar 12 18:23:39 crc kubenswrapper[4926]: E0312 18:23:39.318619 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerName="init" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.318625 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerName="init" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.318788 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" containerName="dnsmasq-dns" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.319774 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.322311 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.333262 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81afb8cd-af2c-4515-b4d1-893903371af0" (UID: "81afb8cd-af2c-4515-b4d1-893903371af0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.346552 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-nb\") pod \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.346735 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwp6h\" (UniqueName: \"kubernetes.io/projected/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-kube-api-access-rwp6h\") pod \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.346857 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0\") pod \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.346899 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-svc\") pod \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.346920 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-sb\") pod \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.346945 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-config\") pod \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347130 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347181 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm8xc\" (UniqueName: \"kubernetes.io/projected/75ab53a0-e811-4b31-8a1e-d71be4115b2c-kube-api-access-sm8xc\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347217 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347242 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-config\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347272 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347333 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347351 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347401 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347418 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347481 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347509 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347530 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpn6w\" (UniqueName: \"kubernetes.io/projected/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-kube-api-access-jpn6w\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347575 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347585 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6qc2\" (UniqueName: \"kubernetes.io/projected/81afb8cd-af2c-4515-b4d1-893903371af0-kube-api-access-v6qc2\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347596 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/81afb8cd-af2c-4515-b4d1-893903371af0-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.347604 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.350127 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.350961 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.351966 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-config\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.352017 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.353044 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.357401 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-kube-api-access-rwp6h" (OuterVolumeSpecName: "kube-api-access-rwp6h") pod "b5cd3f65-2af4-4a38-ab87-c266452f8c5a" (UID: "b5cd3f65-2af4-4a38-ab87-c266452f8c5a"). InnerVolumeSpecName "kube-api-access-rwp6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.367785 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.374179 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.374491 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.376121 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.377810 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.380852 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-scripts\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.392999 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm8xc\" (UniqueName: \"kubernetes.io/projected/75ab53a0-e811-4b31-8a1e-d71be4115b2c-kube-api-access-sm8xc\") pod \"dnsmasq-dns-5c9776ccc5-rln7t\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.404623 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpn6w\" (UniqueName: \"kubernetes.io/projected/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-kube-api-access-jpn6w\") pod \"cinder-scheduler-0\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.441367 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data" (OuterVolumeSpecName: "config-data") pod "81afb8cd-af2c-4515-b4d1-893903371af0" (UID: "81afb8cd-af2c-4515-b4d1-893903371af0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.448008 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5cd3f65-2af4-4a38-ab87-c266452f8c5a" (UID: "b5cd3f65-2af4-4a38-ab87-c266452f8c5a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.448022 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5cd3f65-2af4-4a38-ab87-c266452f8c5a" (UID: "b5cd3f65-2af4-4a38-ab87-c266452f8c5a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.448769 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0\") pod \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\" (UID: \"b5cd3f65-2af4-4a38-ab87-c266452f8c5a\") " Mar 12 18:23:39 crc kubenswrapper[4926]: W0312 18:23:39.448922 4926 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b5cd3f65-2af4-4a38-ab87-c266452f8c5a/volumes/kubernetes.io~configmap/dns-swift-storage-0 Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.448955 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5cd3f65-2af4-4a38-ab87-c266452f8c5a" (UID: "b5cd3f65-2af4-4a38-ab87-c266452f8c5a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.449071 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data-custom\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.449103 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.449120 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-scripts\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.449161 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg65z\" (UniqueName: \"kubernetes.io/projected/78cc2e75-39e6-4148-87a5-022cc3690da8-kube-api-access-jg65z\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.449177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78cc2e75-39e6-4148-87a5-022cc3690da8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.449227 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.449298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cc2e75-39e6-4148-87a5-022cc3690da8-logs\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.450025 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.450057 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.450072 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81afb8cd-af2c-4515-b4d1-893903371af0-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.450084 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwp6h\" (UniqueName: \"kubernetes.io/projected/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-kube-api-access-rwp6h\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.453172 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-config" (OuterVolumeSpecName: "config") pod "b5cd3f65-2af4-4a38-ab87-c266452f8c5a" (UID: "b5cd3f65-2af4-4a38-ab87-c266452f8c5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.464318 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5cd3f65-2af4-4a38-ab87-c266452f8c5a" (UID: "b5cd3f65-2af4-4a38-ab87-c266452f8c5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.469884 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5cd3f65-2af4-4a38-ab87-c266452f8c5a" (UID: "b5cd3f65-2af4-4a38-ab87-c266452f8c5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.551951 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data-custom\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552002 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552021 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-scripts\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552065 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg65z\" (UniqueName: \"kubernetes.io/projected/78cc2e75-39e6-4148-87a5-022cc3690da8-kube-api-access-jg65z\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552084 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78cc2e75-39e6-4148-87a5-022cc3690da8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552123 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552157 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cc2e75-39e6-4148-87a5-022cc3690da8-logs\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552329 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552342 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552351 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5cd3f65-2af4-4a38-ab87-c266452f8c5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552523 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78cc2e75-39e6-4148-87a5-022cc3690da8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.552980 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cc2e75-39e6-4148-87a5-022cc3690da8-logs\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.556328 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.556851 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.561931 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.563190 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data-custom\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.567954 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-scripts\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.571843 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg65z\" (UniqueName: \"kubernetes.io/projected/78cc2e75-39e6-4148-87a5-022cc3690da8-kube-api-access-jg65z\") pod \"cinder-api-0\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.621776 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.663855 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.771625 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerStarted","Data":"d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef"} Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.772073 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-central-agent" containerID="cri-o://3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c" gracePeriod=30 Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.772621 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.772885 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="proxy-httpd" containerID="cri-o://d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef" gracePeriod=30 Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.772931 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="sg-core" containerID="cri-o://a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b" gracePeriod=30 Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.772969 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-notification-agent" containerID="cri-o://aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84" gracePeriod=30 Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.814392 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" event={"ID":"b5cd3f65-2af4-4a38-ab87-c266452f8c5a","Type":"ContainerDied","Data":"a1cf427ff4b305faffde9b50b631a3a735bc0b296fb816c491a872c6626c0d60"} Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.814461 4926 scope.go:117] "RemoveContainer" containerID="3d75e203f5f24817d9aa7e9a6cf9f743a4b6a96bd884d103fc15a23a410323b0" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.814739 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-rvkqm" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.841032 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" event={"ID":"81afb8cd-af2c-4515-b4d1-893903371af0","Type":"ContainerDied","Data":"3adbb909a2691a9c3475b5b8bcf70c23ddd0ca36e2352ef1830b72181f72ded0"} Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.841148 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78dfff8dc9-fgs8s" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.864074 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.725003928 podStartE2EDuration="55.864050804s" podCreationTimestamp="2026-03-12 18:22:44 +0000 UTC" firstStartedPulling="2026-03-12 18:22:45.823072656 +0000 UTC m=+1206.191698979" lastFinishedPulling="2026-03-12 18:23:38.962119522 +0000 UTC m=+1259.330745855" observedRunningTime="2026-03-12 18:23:39.815720347 +0000 UTC m=+1260.184346690" watchObservedRunningTime="2026-03-12 18:23:39.864050804 +0000 UTC m=+1260.232677137" Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.919502 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rvkqm"] Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.976067 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-rvkqm"] Mar 12 18:23:39 crc kubenswrapper[4926]: I0312 18:23:39.979679 4926 scope.go:117] "RemoveContainer" containerID="2f62316dc5afc9b2e89e828a9e035d936c5f1005f1c85a5b2bad886d39f3e886" Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.013730 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-78dfff8dc9-fgs8s"] Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.039249 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-78dfff8dc9-fgs8s"] Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.079734 4926 scope.go:117] "RemoveContainer" containerID="300d112166ede447ec6708957fbc43526b8d0fa27de369019ba8b434782269d0" Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.133851 4926 scope.go:117] "RemoveContainer" containerID="bee09196dcaf75cb5eadfa250f0ca7febdf9ab076b98f039b2a754122b989aef" Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.144471 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.384315 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rln7t"] Mar 12 18:23:40 crc kubenswrapper[4926]: W0312 18:23:40.385384 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ab53a0_e811_4b31_8a1e_d71be4115b2c.slice/crio-2c7e6fa4e701c31af5cac5d50047d5137f377c93ee8e669dc7d858e4f08b2574 WatchSource:0}: Error finding container 2c7e6fa4e701c31af5cac5d50047d5137f377c93ee8e669dc7d858e4f08b2574: Status 404 returned error can't find the container with id 2c7e6fa4e701c31af5cac5d50047d5137f377c93ee8e669dc7d858e4f08b2574 Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.434571 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:40 crc kubenswrapper[4926]: W0312 18:23:40.447862 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78cc2e75_39e6_4148_87a5_022cc3690da8.slice/crio-d2b1d21f05ea6cb158f8a2f7033f48e548508b5ca903be77119e981940c0caa4 WatchSource:0}: Error finding container d2b1d21f05ea6cb158f8a2f7033f48e548508b5ca903be77119e981940c0caa4: Status 404 returned error can't find the container with id d2b1d21f05ea6cb158f8a2f7033f48e548508b5ca903be77119e981940c0caa4 Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.507329 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81afb8cd-af2c-4515-b4d1-893903371af0" path="/var/lib/kubelet/pods/81afb8cd-af2c-4515-b4d1-893903371af0/volumes" Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.508193 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5cd3f65-2af4-4a38-ab87-c266452f8c5a" path="/var/lib/kubelet/pods/b5cd3f65-2af4-4a38-ab87-c266452f8c5a/volumes" Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.885121 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"78cc2e75-39e6-4148-87a5-022cc3690da8","Type":"ContainerStarted","Data":"d2b1d21f05ea6cb158f8a2f7033f48e548508b5ca903be77119e981940c0caa4"} Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.887207 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" event={"ID":"75ab53a0-e811-4b31-8a1e-d71be4115b2c","Type":"ContainerStarted","Data":"ba45c947e1da77d586b46aef916a37581f794d00be599bae5b1f6a31185bfcd3"} Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.887237 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" event={"ID":"75ab53a0-e811-4b31-8a1e-d71be4115b2c","Type":"ContainerStarted","Data":"2c7e6fa4e701c31af5cac5d50047d5137f377c93ee8e669dc7d858e4f08b2574"} Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.901350 4926 generic.go:334] "Generic (PLEG): container finished" podID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerID="d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef" exitCode=0 Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.901383 4926 generic.go:334] "Generic (PLEG): container finished" podID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerID="a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b" exitCode=2 Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.901391 4926 generic.go:334] "Generic (PLEG): container finished" podID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerID="3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c" exitCode=0 Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.901429 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerDied","Data":"d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef"} Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.901467 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerDied","Data":"a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b"} Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.901477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerDied","Data":"3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c"} Mar 12 18:23:40 crc kubenswrapper[4926]: I0312 18:23:40.919266 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c","Type":"ContainerStarted","Data":"b5b79d1f13639d906d19091084d8e1ea42e28328a4dc8a574f851692ee3f9f8c"} Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.241967 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.408349 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.584817 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6fd6447dc6-7dg85" Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.642147 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-9b55c586b-s7wqs"] Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.642354 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-9b55c586b-s7wqs" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api-log" containerID="cri-o://0ae594d417f33a2af41173b2f1260c8120ad7232590b0279e7d83cb0dbe67a9c" gracePeriod=30 Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.642591 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-9b55c586b-s7wqs" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api" containerID="cri-o://8231f1ef7dfab85032bd8e8cc8a2434c72aeff5f8977a2307df4d52b313ce335" gracePeriod=30 Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.958273 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c","Type":"ContainerStarted","Data":"5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc"} Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.983951 4926 generic.go:334] "Generic (PLEG): container finished" podID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerID="0ae594d417f33a2af41173b2f1260c8120ad7232590b0279e7d83cb0dbe67a9c" exitCode=143 Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.984053 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9b55c586b-s7wqs" event={"ID":"2f1f2a42-878e-46c0-bd66-4927a4689299","Type":"ContainerDied","Data":"0ae594d417f33a2af41173b2f1260c8120ad7232590b0279e7d83cb0dbe67a9c"} Mar 12 18:23:41 crc kubenswrapper[4926]: I0312 18:23:41.992131 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"78cc2e75-39e6-4148-87a5-022cc3690da8","Type":"ContainerStarted","Data":"48ea775ffdf8bca7487b51cdb7c8d987f732c9702e1c95ad138a8eebbbab7c90"} Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.000964 4926 generic.go:334] "Generic (PLEG): container finished" podID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerID="ba45c947e1da77d586b46aef916a37581f794d00be599bae5b1f6a31185bfcd3" exitCode=0 Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.002642 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" event={"ID":"75ab53a0-e811-4b31-8a1e-d71be4115b2c","Type":"ContainerDied","Data":"ba45c947e1da77d586b46aef916a37581f794d00be599bae5b1f6a31185bfcd3"} Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.449550 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.773154 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65c5c86775-mct68"] Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.773377 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65c5c86775-mct68" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-api" containerID="cri-o://94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e" gracePeriod=30 Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.773978 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-65c5c86775-mct68" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-httpd" containerID="cri-o://d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217" gracePeriod=30 Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.789479 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7fc6496dbc-7qwrw"] Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.791015 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.808082 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7fc6496dbc-7qwrw"] Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.854327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-httpd-config\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.854422 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-ovndb-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.854465 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-combined-ca-bundle\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.854489 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-config\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.854524 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-internal-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.854557 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kxfv\" (UniqueName: \"kubernetes.io/projected/bf1502ef-50a4-45e0-b193-a6e25abccb32-kube-api-access-4kxfv\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.854610 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-public-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.926211 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-65c5c86775-mct68" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.159:9696/\": read tcp 10.217.0.2:49608->10.217.0.159:9696: read: connection reset by peer" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.956341 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-public-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.956675 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-httpd-config\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.956822 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-ovndb-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.956955 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-combined-ca-bundle\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.957075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-config\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.957692 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-internal-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.957868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kxfv\" (UniqueName: \"kubernetes.io/projected/bf1502ef-50a4-45e0-b193-a6e25abccb32-kube-api-access-4kxfv\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.961317 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-config\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.963240 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-internal-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.963953 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-httpd-config\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.964802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-ovndb-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.965054 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-public-tls-certs\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.978125 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1502ef-50a4-45e0-b193-a6e25abccb32-combined-ca-bundle\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:42 crc kubenswrapper[4926]: I0312 18:23:42.990370 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kxfv\" (UniqueName: \"kubernetes.io/projected/bf1502ef-50a4-45e0-b193-a6e25abccb32-kube-api-access-4kxfv\") pod \"neutron-7fc6496dbc-7qwrw\" (UID: \"bf1502ef-50a4-45e0-b193-a6e25abccb32\") " pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.024686 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c","Type":"ContainerStarted","Data":"433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74"} Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.026581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"78cc2e75-39e6-4148-87a5-022cc3690da8","Type":"ContainerStarted","Data":"a12ed3e178db267afb04649b5718c1d2bcc9a78faa88e5e9519c5b32b47a1362"} Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.026711 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api" containerID="cri-o://a12ed3e178db267afb04649b5718c1d2bcc9a78faa88e5e9519c5b32b47a1362" gracePeriod=30 Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.026732 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.026711 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api-log" containerID="cri-o://48ea775ffdf8bca7487b51cdb7c8d987f732c9702e1c95ad138a8eebbbab7c90" gracePeriod=30 Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.029602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" event={"ID":"75ab53a0-e811-4b31-8a1e-d71be4115b2c","Type":"ContainerStarted","Data":"ad9f00f4a3b221ab329deab372c49c3816c4c0e0f8a0ad87e7aefa0a45c31c5b"} Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.029744 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.037213 4926 generic.go:334] "Generic (PLEG): container finished" podID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerID="d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217" exitCode=0 Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.037281 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c5c86775-mct68" event={"ID":"2beed02e-2edf-4b52-8ea6-ae2dae7502d8","Type":"ContainerDied","Data":"d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217"} Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.044992 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.08737164 podStartE2EDuration="4.044977328s" podCreationTimestamp="2026-03-12 18:23:39 +0000 UTC" firstStartedPulling="2026-03-12 18:23:40.154063974 +0000 UTC m=+1260.522690307" lastFinishedPulling="2026-03-12 18:23:41.111669662 +0000 UTC m=+1261.480295995" observedRunningTime="2026-03-12 18:23:43.041335333 +0000 UTC m=+1263.409961666" watchObservedRunningTime="2026-03-12 18:23:43.044977328 +0000 UTC m=+1263.413603661" Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.066600 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" podStartSLOduration=4.066585025 podStartE2EDuration="4.066585025s" podCreationTimestamp="2026-03-12 18:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:43.061577879 +0000 UTC m=+1263.430204202" watchObservedRunningTime="2026-03-12 18:23:43.066585025 +0000 UTC m=+1263.435211358" Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.082283 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.082265117 podStartE2EDuration="4.082265117s" podCreationTimestamp="2026-03-12 18:23:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:43.081212805 +0000 UTC m=+1263.449839138" watchObservedRunningTime="2026-03-12 18:23:43.082265117 +0000 UTC m=+1263.450891450" Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.172031 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:43 crc kubenswrapper[4926]: I0312 18:23:43.890232 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7fc6496dbc-7qwrw"] Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.070979 4926 generic.go:334] "Generic (PLEG): container finished" podID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerID="a12ed3e178db267afb04649b5718c1d2bcc9a78faa88e5e9519c5b32b47a1362" exitCode=0 Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.071024 4926 generic.go:334] "Generic (PLEG): container finished" podID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerID="48ea775ffdf8bca7487b51cdb7c8d987f732c9702e1c95ad138a8eebbbab7c90" exitCode=143 Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.071065 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"78cc2e75-39e6-4148-87a5-022cc3690da8","Type":"ContainerDied","Data":"a12ed3e178db267afb04649b5718c1d2bcc9a78faa88e5e9519c5b32b47a1362"} Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.071104 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"78cc2e75-39e6-4148-87a5-022cc3690da8","Type":"ContainerDied","Data":"48ea775ffdf8bca7487b51cdb7c8d987f732c9702e1c95ad138a8eebbbab7c90"} Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.071114 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"78cc2e75-39e6-4148-87a5-022cc3690da8","Type":"ContainerDied","Data":"d2b1d21f05ea6cb158f8a2f7033f48e548508b5ca903be77119e981940c0caa4"} Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.071122 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2b1d21f05ea6cb158f8a2f7033f48e548508b5ca903be77119e981940c0caa4" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.083258 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7fc6496dbc-7qwrw" event={"ID":"bf1502ef-50a4-45e0-b193-a6e25abccb32","Type":"ContainerStarted","Data":"37d389d09dc11310aca6959f9e567d011711a2fc227393db818ccf9f61003f62"} Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.111074 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183223 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg65z\" (UniqueName: \"kubernetes.io/projected/78cc2e75-39e6-4148-87a5-022cc3690da8-kube-api-access-jg65z\") pod \"78cc2e75-39e6-4148-87a5-022cc3690da8\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183268 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78cc2e75-39e6-4148-87a5-022cc3690da8-etc-machine-id\") pod \"78cc2e75-39e6-4148-87a5-022cc3690da8\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183322 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data\") pod \"78cc2e75-39e6-4148-87a5-022cc3690da8\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183427 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-scripts\") pod \"78cc2e75-39e6-4148-87a5-022cc3690da8\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183443 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78cc2e75-39e6-4148-87a5-022cc3690da8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "78cc2e75-39e6-4148-87a5-022cc3690da8" (UID: "78cc2e75-39e6-4148-87a5-022cc3690da8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183469 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data-custom\") pod \"78cc2e75-39e6-4148-87a5-022cc3690da8\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183525 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-combined-ca-bundle\") pod \"78cc2e75-39e6-4148-87a5-022cc3690da8\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.183541 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cc2e75-39e6-4148-87a5-022cc3690da8-logs\") pod \"78cc2e75-39e6-4148-87a5-022cc3690da8\" (UID: \"78cc2e75-39e6-4148-87a5-022cc3690da8\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.184050 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78cc2e75-39e6-4148-87a5-022cc3690da8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.187009 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78cc2e75-39e6-4148-87a5-022cc3690da8-logs" (OuterVolumeSpecName: "logs") pod "78cc2e75-39e6-4148-87a5-022cc3690da8" (UID: "78cc2e75-39e6-4148-87a5-022cc3690da8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.216558 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-scripts" (OuterVolumeSpecName: "scripts") pod "78cc2e75-39e6-4148-87a5-022cc3690da8" (UID: "78cc2e75-39e6-4148-87a5-022cc3690da8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.222318 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "78cc2e75-39e6-4148-87a5-022cc3690da8" (UID: "78cc2e75-39e6-4148-87a5-022cc3690da8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.227824 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78cc2e75-39e6-4148-87a5-022cc3690da8-kube-api-access-jg65z" (OuterVolumeSpecName: "kube-api-access-jg65z") pod "78cc2e75-39e6-4148-87a5-022cc3690da8" (UID: "78cc2e75-39e6-4148-87a5-022cc3690da8"). InnerVolumeSpecName "kube-api-access-jg65z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.287592 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.287621 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.287630 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78cc2e75-39e6-4148-87a5-022cc3690da8-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.287638 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg65z\" (UniqueName: \"kubernetes.io/projected/78cc2e75-39e6-4148-87a5-022cc3690da8-kube-api-access-jg65z\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.291803 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78cc2e75-39e6-4148-87a5-022cc3690da8" (UID: "78cc2e75-39e6-4148-87a5-022cc3690da8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.329403 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data" (OuterVolumeSpecName: "config-data") pod "78cc2e75-39e6-4148-87a5-022cc3690da8" (UID: "78cc2e75-39e6-4148-87a5-022cc3690da8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.389630 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.389668 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78cc2e75-39e6-4148-87a5-022cc3690da8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.558815 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.658457 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-65c5c86775-mct68" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.159:9696/\": dial tcp 10.217.0.159:9696: connect: connection refused" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.771266 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.901471 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-combined-ca-bundle\") pod \"5fdfffa4-937c-4167-8545-d34f2007fbc9\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.901753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-sg-core-conf-yaml\") pod \"5fdfffa4-937c-4167-8545-d34f2007fbc9\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.901873 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-log-httpd\") pod \"5fdfffa4-937c-4167-8545-d34f2007fbc9\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.901957 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-scripts\") pod \"5fdfffa4-937c-4167-8545-d34f2007fbc9\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.901984 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-config-data\") pod \"5fdfffa4-937c-4167-8545-d34f2007fbc9\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.902025 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-run-httpd\") pod \"5fdfffa4-937c-4167-8545-d34f2007fbc9\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.902070 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smrs2\" (UniqueName: \"kubernetes.io/projected/5fdfffa4-937c-4167-8545-d34f2007fbc9-kube-api-access-smrs2\") pod \"5fdfffa4-937c-4167-8545-d34f2007fbc9\" (UID: \"5fdfffa4-937c-4167-8545-d34f2007fbc9\") " Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.902385 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5fdfffa4-937c-4167-8545-d34f2007fbc9" (UID: "5fdfffa4-937c-4167-8545-d34f2007fbc9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.902507 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5fdfffa4-937c-4167-8545-d34f2007fbc9" (UID: "5fdfffa4-937c-4167-8545-d34f2007fbc9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.914093 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-scripts" (OuterVolumeSpecName: "scripts") pod "5fdfffa4-937c-4167-8545-d34f2007fbc9" (UID: "5fdfffa4-937c-4167-8545-d34f2007fbc9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.917628 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fdfffa4-937c-4167-8545-d34f2007fbc9-kube-api-access-smrs2" (OuterVolumeSpecName: "kube-api-access-smrs2") pod "5fdfffa4-937c-4167-8545-d34f2007fbc9" (UID: "5fdfffa4-937c-4167-8545-d34f2007fbc9"). InnerVolumeSpecName "kube-api-access-smrs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:44 crc kubenswrapper[4926]: I0312 18:23:44.950097 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5fdfffa4-937c-4167-8545-d34f2007fbc9" (UID: "5fdfffa4-937c-4167-8545-d34f2007fbc9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.006115 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smrs2\" (UniqueName: \"kubernetes.io/projected/5fdfffa4-937c-4167-8545-d34f2007fbc9-kube-api-access-smrs2\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.006150 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.006160 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.006170 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.006178 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5fdfffa4-937c-4167-8545-d34f2007fbc9-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.051808 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fdfffa4-937c-4167-8545-d34f2007fbc9" (UID: "5fdfffa4-937c-4167-8545-d34f2007fbc9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.075745 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-config-data" (OuterVolumeSpecName: "config-data") pod "5fdfffa4-937c-4167-8545-d34f2007fbc9" (UID: "5fdfffa4-937c-4167-8545-d34f2007fbc9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.108744 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.108793 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdfffa4-937c-4167-8545-d34f2007fbc9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.114883 4926 generic.go:334] "Generic (PLEG): container finished" podID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerID="aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84" exitCode=0 Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.114948 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerDied","Data":"aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84"} Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.114976 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5fdfffa4-937c-4167-8545-d34f2007fbc9","Type":"ContainerDied","Data":"fde586bc97eba127841e218475be2e383628bdf6feeb7595fc6e0b0daaf96628"} Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.114992 4926 scope.go:117] "RemoveContainer" containerID="d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.115111 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.135651 4926 generic.go:334] "Generic (PLEG): container finished" podID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerID="8231f1ef7dfab85032bd8e8cc8a2434c72aeff5f8977a2307df4d52b313ce335" exitCode=0 Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.135713 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9b55c586b-s7wqs" event={"ID":"2f1f2a42-878e-46c0-bd66-4927a4689299","Type":"ContainerDied","Data":"8231f1ef7dfab85032bd8e8cc8a2434c72aeff5f8977a2307df4d52b313ce335"} Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.137960 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7fc6496dbc-7qwrw" event={"ID":"bf1502ef-50a4-45e0-b193-a6e25abccb32","Type":"ContainerStarted","Data":"35a613e6b532db3489cd114f2dff0c75f70bff1b029f6056046d9fba6744021e"} Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.138019 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7fc6496dbc-7qwrw" event={"ID":"bf1502ef-50a4-45e0-b193-a6e25abccb32","Type":"ContainerStarted","Data":"0867465e0603cad4755aa2a9a92b705554f77ad30ccd6ba55d0ec179f8c1aab3"} Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.138054 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.138129 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.171046 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7fc6496dbc-7qwrw" podStartSLOduration=3.17102843 podStartE2EDuration="3.17102843s" podCreationTimestamp="2026-03-12 18:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:45.169873295 +0000 UTC m=+1265.538499628" watchObservedRunningTime="2026-03-12 18:23:45.17102843 +0000 UTC m=+1265.539654763" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.211851 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.211988 4926 scope.go:117] "RemoveContainer" containerID="a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.220789 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.238655 4926 scope.go:117] "RemoveContainer" containerID="aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.247167 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.260465 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.272284 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.272830 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api-log" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.272856 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api-log" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.272872 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-central-agent" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.272879 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-central-agent" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.272893 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.272899 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.272908 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="proxy-httpd" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.272914 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="proxy-httpd" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.272922 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-notification-agent" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.272932 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-notification-agent" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.272959 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="sg-core" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.272967 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="sg-core" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.273173 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api-log" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.273192 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-central-agent" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.273207 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="sg-core" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.273218 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" containerName="cinder-api" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.273234 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="ceilometer-notification-agent" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.273252 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" containerName="proxy-httpd" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.275403 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.281430 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.281578 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.281673 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.290920 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.294815 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.296783 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.297203 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.298311 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.299676 4926 scope.go:117] "RemoveContainer" containerID="3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.300188 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.333664 4926 scope.go:117] "RemoveContainer" containerID="d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.334418 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef\": container with ID starting with d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef not found: ID does not exist" containerID="d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.334463 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef"} err="failed to get container status \"d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef\": rpc error: code = NotFound desc = could not find container \"d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef\": container with ID starting with d6ab048172eca2cb50eb57623ff1095082421f1f4c5137f11d5c1497dd206bef not found: ID does not exist" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.334485 4926 scope.go:117] "RemoveContainer" containerID="a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.335356 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b\": container with ID starting with a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b not found: ID does not exist" containerID="a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.335379 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b"} err="failed to get container status \"a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b\": rpc error: code = NotFound desc = could not find container \"a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b\": container with ID starting with a28f9bce7474ed0af2327b3b39f88ffd436344c5e5a64a5327965e743ea69f3b not found: ID does not exist" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.335394 4926 scope.go:117] "RemoveContainer" containerID="aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.336292 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84\": container with ID starting with aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84 not found: ID does not exist" containerID="aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.336320 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84"} err="failed to get container status \"aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84\": rpc error: code = NotFound desc = could not find container \"aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84\": container with ID starting with aff7f75f8e01b74292a231c7e54734896e7fa5d7cb06847602f5501bd5510f84 not found: ID does not exist" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.336338 4926 scope.go:117] "RemoveContainer" containerID="3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c" Mar 12 18:23:45 crc kubenswrapper[4926]: E0312 18:23:45.337539 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c\": container with ID starting with 3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c not found: ID does not exist" containerID="3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.337564 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c"} err="failed to get container status \"3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c\": rpc error: code = NotFound desc = could not find container \"3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c\": container with ID starting with 3d2006e37fc8315276f34e264200f53f1fbfd12d88a77b340db54bf40d377b3c not found: ID does not exist" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418375 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bmcf\" (UniqueName: \"kubernetes.io/projected/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-kube-api-access-8bmcf\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418411 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-run-httpd\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418464 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-config-data-custom\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418479 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418499 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-config-data\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418643 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418811 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418843 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd14509-9e82-40a2-aea4-c6ad4250be05-logs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418878 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd14509-9e82-40a2-aea4-c6ad4250be05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418913 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-scripts\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.418987 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnv8k\" (UniqueName: \"kubernetes.io/projected/edd14509-9e82-40a2-aea4-c6ad4250be05-kube-api-access-lnv8k\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.419019 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-log-httpd\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.419034 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-scripts\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.419111 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-config-data\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.419162 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.419202 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.475660 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-config-data-custom\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524418 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524460 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-config-data\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524496 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524525 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd14509-9e82-40a2-aea4-c6ad4250be05-logs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524558 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd14509-9e82-40a2-aea4-c6ad4250be05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524576 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-scripts\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524609 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnv8k\" (UniqueName: \"kubernetes.io/projected/edd14509-9e82-40a2-aea4-c6ad4250be05-kube-api-access-lnv8k\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-scripts\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524640 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-log-httpd\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524672 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-config-data\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524690 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524745 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bmcf\" (UniqueName: \"kubernetes.io/projected/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-kube-api-access-8bmcf\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.524764 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-run-httpd\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.525282 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-run-httpd\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.531692 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd14509-9e82-40a2-aea4-c6ad4250be05-logs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.532601 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-log-httpd\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.533066 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.533206 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/edd14509-9e82-40a2-aea4-c6ad4250be05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.536338 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-config-data\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.536757 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-scripts\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.537031 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.537197 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-config-data-custom\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.537302 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-scripts\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.537413 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.538138 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.547619 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-config-data\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.548686 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd14509-9e82-40a2-aea4-c6ad4250be05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.550734 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bmcf\" (UniqueName: \"kubernetes.io/projected/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-kube-api-access-8bmcf\") pod \"ceilometer-0\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.551141 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnv8k\" (UniqueName: \"kubernetes.io/projected/edd14509-9e82-40a2-aea4-c6ad4250be05-kube-api-access-lnv8k\") pod \"cinder-api-0\" (UID: \"edd14509-9e82-40a2-aea4-c6ad4250be05\") " pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.626151 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qldmj\" (UniqueName: \"kubernetes.io/projected/2f1f2a42-878e-46c0-bd66-4927a4689299-kube-api-access-qldmj\") pod \"2f1f2a42-878e-46c0-bd66-4927a4689299\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.626206 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data\") pod \"2f1f2a42-878e-46c0-bd66-4927a4689299\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.626264 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data-custom\") pod \"2f1f2a42-878e-46c0-bd66-4927a4689299\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.626298 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-combined-ca-bundle\") pod \"2f1f2a42-878e-46c0-bd66-4927a4689299\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.626384 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1f2a42-878e-46c0-bd66-4927a4689299-logs\") pod \"2f1f2a42-878e-46c0-bd66-4927a4689299\" (UID: \"2f1f2a42-878e-46c0-bd66-4927a4689299\") " Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.627417 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f1f2a42-878e-46c0-bd66-4927a4689299-logs" (OuterVolumeSpecName: "logs") pod "2f1f2a42-878e-46c0-bd66-4927a4689299" (UID: "2f1f2a42-878e-46c0-bd66-4927a4689299"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.630267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2f1f2a42-878e-46c0-bd66-4927a4689299" (UID: "2f1f2a42-878e-46c0-bd66-4927a4689299"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.633011 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.637657 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f1f2a42-878e-46c0-bd66-4927a4689299-kube-api-access-qldmj" (OuterVolumeSpecName: "kube-api-access-qldmj") pod "2f1f2a42-878e-46c0-bd66-4927a4689299" (UID: "2f1f2a42-878e-46c0-bd66-4927a4689299"). InnerVolumeSpecName "kube-api-access-qldmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.663055 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f1f2a42-878e-46c0-bd66-4927a4689299" (UID: "2f1f2a42-878e-46c0-bd66-4927a4689299"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.673901 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.696695 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data" (OuterVolumeSpecName: "config-data") pod "2f1f2a42-878e-46c0-bd66-4927a4689299" (UID: "2f1f2a42-878e-46c0-bd66-4927a4689299"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.728581 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f1f2a42-878e-46c0-bd66-4927a4689299-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.728614 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qldmj\" (UniqueName: \"kubernetes.io/projected/2f1f2a42-878e-46c0-bd66-4927a4689299-kube-api-access-qldmj\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.728625 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.728636 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:45 crc kubenswrapper[4926]: I0312 18:23:45.728646 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f1f2a42-878e-46c0-bd66-4927a4689299-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.151165 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9b55c586b-s7wqs" event={"ID":"2f1f2a42-878e-46c0-bd66-4927a4689299","Type":"ContainerDied","Data":"28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc"} Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.151157 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9b55c586b-s7wqs" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.151450 4926 scope.go:117] "RemoveContainer" containerID="8231f1ef7dfab85032bd8e8cc8a2434c72aeff5f8977a2307df4d52b313ce335" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.204989 4926 scope.go:117] "RemoveContainer" containerID="0ae594d417f33a2af41173b2f1260c8120ad7232590b0279e7d83cb0dbe67a9c" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.216618 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-9b55c586b-s7wqs"] Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.228867 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-9b55c586b-s7wqs"] Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.265431 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.279717 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.505944 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" path="/var/lib/kubelet/pods/2f1f2a42-878e-46c0-bd66-4927a4689299/volumes" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.506595 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fdfffa4-937c-4167-8545-d34f2007fbc9" path="/var/lib/kubelet/pods/5fdfffa4-937c-4167-8545-d34f2007fbc9/volumes" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.509559 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78cc2e75-39e6-4148-87a5-022cc3690da8" path="/var/lib/kubelet/pods/78cc2e75-39e6-4148-87a5-022cc3690da8/volumes" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.562175 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:23:46 crc kubenswrapper[4926]: I0312 18:23:46.602280 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:23:47 crc kubenswrapper[4926]: I0312 18:23:47.167865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerStarted","Data":"f1898b8c34b4b06f2e10cb4f43fffdb55131049338d63dbe39254ab53b9a2db0"} Mar 12 18:23:47 crc kubenswrapper[4926]: I0312 18:23:47.170306 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edd14509-9e82-40a2-aea4-c6ad4250be05","Type":"ContainerStarted","Data":"e10b6da3dd1ff8440bdbc1a8379a0386e7115d448c1e89e355b494fc9b57f163"} Mar 12 18:23:47 crc kubenswrapper[4926]: I0312 18:23:47.170329 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edd14509-9e82-40a2-aea4-c6ad4250be05","Type":"ContainerStarted","Data":"deaafe388600bb6e713e42d64fd7e99daa40243a141341603a4a24f22c85d95b"} Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.190568 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"edd14509-9e82-40a2-aea4-c6ad4250be05","Type":"ContainerStarted","Data":"72ce98f1d314783c0caefbbcaa07ad14bc6f88c8f23b466acb4769e4fb57365f"} Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.190942 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.196391 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerStarted","Data":"a367820b72291dce3847ddbcc2b3a9f8493b17ddb9b80ab37f0c0b123bd5c72e"} Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.196453 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerStarted","Data":"a9eba08d9bcde00929ca468e28634454c84f19ace5c5074a2b2f0456e3403fef"} Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.216630 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.216611607 podStartE2EDuration="3.216611607s" podCreationTimestamp="2026-03-12 18:23:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:48.213096097 +0000 UTC m=+1268.581722440" watchObservedRunningTime="2026-03-12 18:23:48.216611607 +0000 UTC m=+1268.585237940" Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.310642 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5c6848d8cd-cq57n" Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.369974 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-89554fb64-s9c6q"] Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.370208 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon-log" containerID="cri-o://a41e698428edf8a4dccc211b474a012677761bd1a52d186e37df8c3d1445ee62" gracePeriod=30 Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.370681 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" containerID="cri-o://2388379aefe6f54987e9387c0ab60e55776d5a81c5e2c5b0f21608b48b6e8fa5" gracePeriod=30 Mar 12 18:23:48 crc kubenswrapper[4926]: I0312 18:23:48.383030 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.048753 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.108826 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-httpd-config\") pod \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.108949 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-combined-ca-bundle\") pod \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.109148 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-ovndb-tls-certs\") pod \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.109656 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-internal-tls-certs\") pod \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.110375 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzgvh\" (UniqueName: \"kubernetes.io/projected/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-kube-api-access-nzgvh\") pod \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.110416 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-config\") pod \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.110495 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-public-tls-certs\") pod \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\" (UID: \"2beed02e-2edf-4b52-8ea6-ae2dae7502d8\") " Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.114030 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2beed02e-2edf-4b52-8ea6-ae2dae7502d8" (UID: "2beed02e-2edf-4b52-8ea6-ae2dae7502d8"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.114453 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-kube-api-access-nzgvh" (OuterVolumeSpecName: "kube-api-access-nzgvh") pod "2beed02e-2edf-4b52-8ea6-ae2dae7502d8" (UID: "2beed02e-2edf-4b52-8ea6-ae2dae7502d8"). InnerVolumeSpecName "kube-api-access-nzgvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.175796 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2beed02e-2edf-4b52-8ea6-ae2dae7502d8" (UID: "2beed02e-2edf-4b52-8ea6-ae2dae7502d8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.187977 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2beed02e-2edf-4b52-8ea6-ae2dae7502d8" (UID: "2beed02e-2edf-4b52-8ea6-ae2dae7502d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.202505 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2beed02e-2edf-4b52-8ea6-ae2dae7502d8" (UID: "2beed02e-2edf-4b52-8ea6-ae2dae7502d8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.209400 4926 generic.go:334] "Generic (PLEG): container finished" podID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerID="94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e" exitCode=0 Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.209475 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c5c86775-mct68" event={"ID":"2beed02e-2edf-4b52-8ea6-ae2dae7502d8","Type":"ContainerDied","Data":"94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e"} Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.209502 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-65c5c86775-mct68" event={"ID":"2beed02e-2edf-4b52-8ea6-ae2dae7502d8","Type":"ContainerDied","Data":"7b5dc2f13fd79b24bbd53ad5c8577f6fcd9167f1b3fea4e0a8139bb248f187ac"} Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.209518 4926 scope.go:117] "RemoveContainer" containerID="d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.209678 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-65c5c86775-mct68" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.213028 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.213060 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.213068 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.213079 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.213090 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzgvh\" (UniqueName: \"kubernetes.io/projected/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-kube-api-access-nzgvh\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.219101 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerStarted","Data":"8a7449dd308bb99ab10627007eaa35755447eacebb3f5175eab4f716527295a3"} Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.228281 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-config" (OuterVolumeSpecName: "config") pod "2beed02e-2edf-4b52-8ea6-ae2dae7502d8" (UID: "2beed02e-2edf-4b52-8ea6-ae2dae7502d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.233943 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2beed02e-2edf-4b52-8ea6-ae2dae7502d8" (UID: "2beed02e-2edf-4b52-8ea6-ae2dae7502d8"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.296192 4926 scope.go:117] "RemoveContainer" containerID="94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.315322 4926 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.315374 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2beed02e-2edf-4b52-8ea6-ae2dae7502d8-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.319742 4926 scope.go:117] "RemoveContainer" containerID="d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217" Mar 12 18:23:49 crc kubenswrapper[4926]: E0312 18:23:49.320160 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217\": container with ID starting with d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217 not found: ID does not exist" containerID="d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.320188 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217"} err="failed to get container status \"d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217\": rpc error: code = NotFound desc = could not find container \"d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217\": container with ID starting with d88e7111c6324f8b6f31f6801ff8076b863ecd2243e9702f347b8ae37763d217 not found: ID does not exist" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.320212 4926 scope.go:117] "RemoveContainer" containerID="94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e" Mar 12 18:23:49 crc kubenswrapper[4926]: E0312 18:23:49.320525 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e\": container with ID starting with 94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e not found: ID does not exist" containerID="94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.320549 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e"} err="failed to get container status \"94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e\": rpc error: code = NotFound desc = could not find container \"94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e\": container with ID starting with 94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e not found: ID does not exist" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.555738 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-65c5c86775-mct68"] Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.564795 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-65c5c86775-mct68"] Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.624690 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.690206 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-gbmf8"] Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.694610 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" podUID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerName="dnsmasq-dns" containerID="cri-o://3e365f4cb50835b1180ff7c19d7fc397df5fff675446efdb00f5237496ddec2b" gracePeriod=10 Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.870228 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.916372 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:49 crc kubenswrapper[4926]: I0312 18:23:49.976001 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.210069 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.234048 4926 generic.go:334] "Generic (PLEG): container finished" podID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerID="3e365f4cb50835b1180ff7c19d7fc397df5fff675446efdb00f5237496ddec2b" exitCode=0 Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.234139 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" event={"ID":"08abfe56-0e5c-4634-9a1a-488e2bbb587d","Type":"ContainerDied","Data":"3e365f4cb50835b1180ff7c19d7fc397df5fff675446efdb00f5237496ddec2b"} Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.234183 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" event={"ID":"08abfe56-0e5c-4634-9a1a-488e2bbb587d","Type":"ContainerDied","Data":"07e4278fe40ae8695fb94d6a2d5bec78c86b69a3255ccbd19aa22a7865e54876"} Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.234198 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07e4278fe40ae8695fb94d6a2d5bec78c86b69a3255ccbd19aa22a7865e54876" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.234263 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="cinder-scheduler" containerID="cri-o://5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc" gracePeriod=30 Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.234611 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="probe" containerID="cri-o://433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74" gracePeriod=30 Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.263740 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.334721 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxtxm\" (UniqueName: \"kubernetes.io/projected/08abfe56-0e5c-4634-9a1a-488e2bbb587d-kube-api-access-pxtxm\") pod \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.334837 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-config\") pod \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.334959 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-sb\") pod \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.335005 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-svc\") pod \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.335099 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-swift-storage-0\") pod \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.335130 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-nb\") pod \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\" (UID: \"08abfe56-0e5c-4634-9a1a-488e2bbb587d\") " Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.340587 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08abfe56-0e5c-4634-9a1a-488e2bbb587d-kube-api-access-pxtxm" (OuterVolumeSpecName: "kube-api-access-pxtxm") pod "08abfe56-0e5c-4634-9a1a-488e2bbb587d" (UID: "08abfe56-0e5c-4634-9a1a-488e2bbb587d"). InnerVolumeSpecName "kube-api-access-pxtxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.391797 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08abfe56-0e5c-4634-9a1a-488e2bbb587d" (UID: "08abfe56-0e5c-4634-9a1a-488e2bbb587d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.392685 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-config" (OuterVolumeSpecName: "config") pod "08abfe56-0e5c-4634-9a1a-488e2bbb587d" (UID: "08abfe56-0e5c-4634-9a1a-488e2bbb587d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.396033 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "08abfe56-0e5c-4634-9a1a-488e2bbb587d" (UID: "08abfe56-0e5c-4634-9a1a-488e2bbb587d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.414025 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "08abfe56-0e5c-4634-9a1a-488e2bbb587d" (UID: "08abfe56-0e5c-4634-9a1a-488e2bbb587d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.439574 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.439650 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.439665 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.439677 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxtxm\" (UniqueName: \"kubernetes.io/projected/08abfe56-0e5c-4634-9a1a-488e2bbb587d-kube-api-access-pxtxm\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.439711 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.485399 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6fdd856968-nmxxn"] Mar 12 18:23:50 crc kubenswrapper[4926]: E0312 18:23:50.485786 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerName="init" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.485801 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerName="init" Mar 12 18:23:50 crc kubenswrapper[4926]: E0312 18:23:50.485817 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api-log" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.485823 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api-log" Mar 12 18:23:50 crc kubenswrapper[4926]: E0312 18:23:50.485834 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-api" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.485840 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-api" Mar 12 18:23:50 crc kubenswrapper[4926]: E0312 18:23:50.485853 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerName="dnsmasq-dns" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.485859 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerName="dnsmasq-dns" Mar 12 18:23:50 crc kubenswrapper[4926]: E0312 18:23:50.485868 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-httpd" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.485873 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-httpd" Mar 12 18:23:50 crc kubenswrapper[4926]: E0312 18:23:50.485888 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.485894 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.486054 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-api" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.486066 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" containerName="neutron-httpd" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.486074 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api-log" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.486085 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" containerName="dnsmasq-dns" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.486095 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f1f2a42-878e-46c0-bd66-4927a4689299" containerName="barbican-api" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.486975 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.502288 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2beed02e-2edf-4b52-8ea6-ae2dae7502d8" path="/var/lib/kubelet/pods/2beed02e-2edf-4b52-8ea6-ae2dae7502d8/volumes" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.508420 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "08abfe56-0e5c-4634-9a1a-488e2bbb587d" (UID: "08abfe56-0e5c-4634-9a1a-488e2bbb587d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.514291 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fdd856968-nmxxn"] Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542669 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-internal-tls-certs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542811 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s94wd\" (UniqueName: \"kubernetes.io/projected/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-kube-api-access-s94wd\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542848 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-logs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542874 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-config-data\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542895 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-combined-ca-bundle\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542914 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-scripts\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542928 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-public-tls-certs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.542988 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08abfe56-0e5c-4634-9a1a-488e2bbb587d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.644286 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s94wd\" (UniqueName: \"kubernetes.io/projected/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-kube-api-access-s94wd\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.644365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-logs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.644397 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-config-data\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.644447 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-combined-ca-bundle\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.644474 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-scripts\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.644496 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-public-tls-certs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.644549 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-internal-tls-certs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.645189 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-logs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.647911 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-scripts\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.648970 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-internal-tls-certs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.649110 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-combined-ca-bundle\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.650082 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-config-data\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.654212 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-public-tls-certs\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.665099 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s94wd\" (UniqueName: \"kubernetes.io/projected/0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81-kube-api-access-s94wd\") pod \"placement-6fdd856968-nmxxn\" (UID: \"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81\") " pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:50 crc kubenswrapper[4926]: I0312 18:23:50.819872 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:51 crc kubenswrapper[4926]: I0312 18:23:51.250535 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerID="433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74" exitCode=0 Mar 12 18:23:51 crc kubenswrapper[4926]: I0312 18:23:51.251094 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-gbmf8" Mar 12 18:23:51 crc kubenswrapper[4926]: I0312 18:23:51.250652 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c","Type":"ContainerDied","Data":"433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74"} Mar 12 18:23:51 crc kubenswrapper[4926]: I0312 18:23:51.319163 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fdd856968-nmxxn"] Mar 12 18:23:51 crc kubenswrapper[4926]: I0312 18:23:51.357675 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-gbmf8"] Mar 12 18:23:51 crc kubenswrapper[4926]: I0312 18:23:51.368691 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-gbmf8"] Mar 12 18:23:51 crc kubenswrapper[4926]: I0312 18:23:51.839320 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:59182->10.217.0.152:8443: read: connection reset by peer" Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.259988 4926 generic.go:334] "Generic (PLEG): container finished" podID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerID="2388379aefe6f54987e9387c0ab60e55776d5a81c5e2c5b0f21608b48b6e8fa5" exitCode=0 Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.260071 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-89554fb64-s9c6q" event={"ID":"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572","Type":"ContainerDied","Data":"2388379aefe6f54987e9387c0ab60e55776d5a81c5e2c5b0f21608b48b6e8fa5"} Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.261399 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fdd856968-nmxxn" event={"ID":"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81","Type":"ContainerStarted","Data":"16d5e379b276abe6132ccec2a9f5cb3df0aa08c2f32851e225b26b48de0a209e"} Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.261535 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fdd856968-nmxxn" event={"ID":"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81","Type":"ContainerStarted","Data":"2030315c68b9ec5e49c8e54e4477a702ee63c1c1a1199c8231f96ac1189cd4d4"} Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.261588 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fdd856968-nmxxn" event={"ID":"0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81","Type":"ContainerStarted","Data":"2975e38110c330439c82a74fcdb5b43027e0eb2466522cad66f4767299630c42"} Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.261772 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.261807 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.263771 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerStarted","Data":"27c0e5b7f4011faa0c2b0ee4f54cfe990c1a5e30f0bbe162914f8380a857f6ef"} Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.263940 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.283015 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6fdd856968-nmxxn" podStartSLOduration=2.282994966 podStartE2EDuration="2.282994966s" podCreationTimestamp="2026-03-12 18:23:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:52.279542968 +0000 UTC m=+1272.648169311" watchObservedRunningTime="2026-03-12 18:23:52.282994966 +0000 UTC m=+1272.651621299" Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.308825 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.471354642 podStartE2EDuration="7.308801256s" podCreationTimestamp="2026-03-12 18:23:45 +0000 UTC" firstStartedPulling="2026-03-12 18:23:46.30459024 +0000 UTC m=+1266.673216573" lastFinishedPulling="2026-03-12 18:23:51.142036854 +0000 UTC m=+1271.510663187" observedRunningTime="2026-03-12 18:23:52.303313623 +0000 UTC m=+1272.671939956" watchObservedRunningTime="2026-03-12 18:23:52.308801256 +0000 UTC m=+1272.677427579" Mar 12 18:23:52 crc kubenswrapper[4926]: I0312 18:23:52.500238 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08abfe56-0e5c-4634-9a1a-488e2bbb587d" path="/var/lib/kubelet/pods/08abfe56-0e5c-4634-9a1a-488e2bbb587d/volumes" Mar 12 18:23:53 crc kubenswrapper[4926]: I0312 18:23:53.217671 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.282033 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerID="5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc" exitCode=0 Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.282112 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c","Type":"ContainerDied","Data":"5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc"} Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.685554 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.736732 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-scripts\") pod \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.736806 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data-custom\") pod \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.736992 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpn6w\" (UniqueName: \"kubernetes.io/projected/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-kube-api-access-jpn6w\") pod \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.737058 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-etc-machine-id\") pod \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.737139 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-combined-ca-bundle\") pod \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.737175 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data\") pod \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\" (UID: \"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c\") " Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.739906 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" (UID: "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.765542 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-kube-api-access-jpn6w" (OuterVolumeSpecName: "kube-api-access-jpn6w") pod "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" (UID: "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c"). InnerVolumeSpecName "kube-api-access-jpn6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.765613 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-scripts" (OuterVolumeSpecName: "scripts") pod "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" (UID: "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.784583 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" (UID: "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.817156 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" (UID: "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.840039 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpn6w\" (UniqueName: \"kubernetes.io/projected/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-kube-api-access-jpn6w\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.840072 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.840082 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.840093 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.840104 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.885114 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data" (OuterVolumeSpecName: "config-data") pod "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" (UID: "1a13f5e0-72f6-4c47-a5ea-349c2d618d8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:23:54 crc kubenswrapper[4926]: I0312 18:23:54.941861 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.292705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1a13f5e0-72f6-4c47-a5ea-349c2d618d8c","Type":"ContainerDied","Data":"b5b79d1f13639d906d19091084d8e1ea42e28328a4dc8a574f851692ee3f9f8c"} Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.292768 4926 scope.go:117] "RemoveContainer" containerID="433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.292815 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.349106 4926 scope.go:117] "RemoveContainer" containerID="5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.355553 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.375423 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.383906 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:55 crc kubenswrapper[4926]: E0312 18:23:55.384367 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="cinder-scheduler" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.384385 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="cinder-scheduler" Mar 12 18:23:55 crc kubenswrapper[4926]: E0312 18:23:55.384417 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="probe" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.384426 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="probe" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.384693 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="probe" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.384719 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" containerName="cinder-scheduler" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.385762 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.387745 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.397261 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.450137 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.450204 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b586d98-12e7-4814-82c6-6724e1b35a77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.450241 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.450274 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktwx7\" (UniqueName: \"kubernetes.io/projected/7b586d98-12e7-4814-82c6-6724e1b35a77-kube-api-access-ktwx7\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.450293 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.450704 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.552557 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.553259 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.553293 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b586d98-12e7-4814-82c6-6724e1b35a77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.553352 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7b586d98-12e7-4814-82c6-6724e1b35a77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.553384 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.553499 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktwx7\" (UniqueName: \"kubernetes.io/projected/7b586d98-12e7-4814-82c6-6724e1b35a77-kube-api-access-ktwx7\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.553534 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.558976 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.567093 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-scripts\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.569254 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-config-data\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.569916 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b586d98-12e7-4814-82c6-6724e1b35a77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.570943 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktwx7\" (UniqueName: \"kubernetes.io/projected/7b586d98-12e7-4814-82c6-6724e1b35a77-kube-api-access-ktwx7\") pod \"cinder-scheduler-0\" (UID: \"7b586d98-12e7-4814-82c6-6724e1b35a77\") " pod="openstack/cinder-scheduler-0" Mar 12 18:23:55 crc kubenswrapper[4926]: I0312 18:23:55.707967 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 12 18:23:56 crc kubenswrapper[4926]: I0312 18:23:56.220452 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 12 18:23:56 crc kubenswrapper[4926]: I0312 18:23:56.314614 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b586d98-12e7-4814-82c6-6724e1b35a77","Type":"ContainerStarted","Data":"b092141ebd6c3282eb64208c6d0eada485cc3f4e87f9e320f3e0527985ac5d9b"} Mar 12 18:23:56 crc kubenswrapper[4926]: I0312 18:23:56.500068 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a13f5e0-72f6-4c47-a5ea-349c2d618d8c" path="/var/lib/kubelet/pods/1a13f5e0-72f6-4c47-a5ea-349c2d618d8c/volumes" Mar 12 18:23:56 crc kubenswrapper[4926]: I0312 18:23:56.817515 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:23:56 crc kubenswrapper[4926]: I0312 18:23:56.817571 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:23:57 crc kubenswrapper[4926]: I0312 18:23:57.328976 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b586d98-12e7-4814-82c6-6724e1b35a77","Type":"ContainerStarted","Data":"88264eae8cfce90c518596add7ceb9977069fc99141c87436e9831d2980346ae"} Mar 12 18:23:57 crc kubenswrapper[4926]: I0312 18:23:57.388025 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5b667c464b-fk8sc" Mar 12 18:23:57 crc kubenswrapper[4926]: I0312 18:23:57.745250 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 12 18:23:58 crc kubenswrapper[4926]: I0312 18:23:58.340584 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7b586d98-12e7-4814-82c6-6724e1b35a77","Type":"ContainerStarted","Data":"3c43768b9e52861674aa32825296172a82245451794d4d7809a7351482d03074"} Mar 12 18:23:58 crc kubenswrapper[4926]: I0312 18:23:58.372106 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.372083915 podStartE2EDuration="3.372083915s" podCreationTimestamp="2026-03-12 18:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:23:58.356927939 +0000 UTC m=+1278.725554292" watchObservedRunningTime="2026-03-12 18:23:58.372083915 +0000 UTC m=+1278.740710248" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.143595 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555664-xqlj7"] Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.144719 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555664-xqlj7" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.149196 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.150390 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.150623 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.173350 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555664-xqlj7"] Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.257907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gznnp\" (UniqueName: \"kubernetes.io/projected/d0a93b50-2038-4bf8-8c5f-bc77148d55f8-kube-api-access-gznnp\") pod \"auto-csr-approver-29555664-xqlj7\" (UID: \"d0a93b50-2038-4bf8-8c5f-bc77148d55f8\") " pod="openshift-infra/auto-csr-approver-29555664-xqlj7" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.359860 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gznnp\" (UniqueName: \"kubernetes.io/projected/d0a93b50-2038-4bf8-8c5f-bc77148d55f8-kube-api-access-gznnp\") pod \"auto-csr-approver-29555664-xqlj7\" (UID: \"d0a93b50-2038-4bf8-8c5f-bc77148d55f8\") " pod="openshift-infra/auto-csr-approver-29555664-xqlj7" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.384498 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gznnp\" (UniqueName: \"kubernetes.io/projected/d0a93b50-2038-4bf8-8c5f-bc77148d55f8-kube-api-access-gznnp\") pod \"auto-csr-approver-29555664-xqlj7\" (UID: \"d0a93b50-2038-4bf8-8c5f-bc77148d55f8\") " pod="openshift-infra/auto-csr-approver-29555664-xqlj7" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.476641 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555664-xqlj7" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.708226 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.922113 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-bbdc94bc7-pmtmm"] Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.925248 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.932427 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.932621 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.932699 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.943322 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bbdc94bc7-pmtmm"] Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971078 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-config-data\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971164 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-public-tls-certs\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971198 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-combined-ca-bundle\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971236 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5xmq\" (UniqueName: \"kubernetes.io/projected/09eb7b3b-c5af-4625-8f1a-83766550711c-kube-api-access-x5xmq\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971329 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/09eb7b3b-c5af-4625-8f1a-83766550711c-etc-swift\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971467 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09eb7b3b-c5af-4625-8f1a-83766550711c-log-httpd\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971583 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-internal-tls-certs\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.971614 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09eb7b3b-c5af-4625-8f1a-83766550711c-run-httpd\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:00 crc kubenswrapper[4926]: I0312 18:24:00.998384 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555664-xqlj7"] Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.072691 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-internal-tls-certs\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.072759 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09eb7b3b-c5af-4625-8f1a-83766550711c-run-httpd\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.072817 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-config-data\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.072875 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-public-tls-certs\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.072906 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-combined-ca-bundle\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.072949 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5xmq\" (UniqueName: \"kubernetes.io/projected/09eb7b3b-c5af-4625-8f1a-83766550711c-kube-api-access-x5xmq\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.072986 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/09eb7b3b-c5af-4625-8f1a-83766550711c-etc-swift\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.073032 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09eb7b3b-c5af-4625-8f1a-83766550711c-log-httpd\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.073265 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09eb7b3b-c5af-4625-8f1a-83766550711c-run-httpd\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.073561 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/09eb7b3b-c5af-4625-8f1a-83766550711c-log-httpd\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.089535 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-internal-tls-certs\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.091205 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/09eb7b3b-c5af-4625-8f1a-83766550711c-etc-swift\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.093486 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-public-tls-certs\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.094321 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-combined-ca-bundle\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.099525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5xmq\" (UniqueName: \"kubernetes.io/projected/09eb7b3b-c5af-4625-8f1a-83766550711c-kube-api-access-x5xmq\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.099951 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09eb7b3b-c5af-4625-8f1a-83766550711c-config-data\") pod \"swift-proxy-bbdc94bc7-pmtmm\" (UID: \"09eb7b3b-c5af-4625-8f1a-83766550711c\") " pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.250820 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.369481 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555664-xqlj7" event={"ID":"d0a93b50-2038-4bf8-8c5f-bc77148d55f8","Type":"ContainerStarted","Data":"b73f746aeaacae651cc800c031298955a2feeae1651f82a533f65b94b8415760"} Mar 12 18:24:01 crc kubenswrapper[4926]: I0312 18:24:01.820801 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bbdc94bc7-pmtmm"] Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.049499 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.051141 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.053409 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-82l47" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.059017 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.059313 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.059495 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.064712 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.094970 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da936793-13b1-4815-a1ec-4d5d609ca5e3-logs\") pod \"da936793-13b1-4815-a1ec-4d5d609ca5e3\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095047 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data-custom\") pod \"da936793-13b1-4815-a1ec-4d5d609ca5e3\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095144 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m7fh\" (UniqueName: \"kubernetes.io/projected/da936793-13b1-4815-a1ec-4d5d609ca5e3-kube-api-access-2m7fh\") pod \"da936793-13b1-4815-a1ec-4d5d609ca5e3\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095215 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-combined-ca-bundle\") pod \"da936793-13b1-4815-a1ec-4d5d609ca5e3\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095270 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data\") pod \"da936793-13b1-4815-a1ec-4d5d609ca5e3\" (UID: \"da936793-13b1-4815-a1ec-4d5d609ca5e3\") " Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095477 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ac368f3-42fb-4f4a-ba68-1686386b017e-openstack-config-secret\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095545 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8ac368f3-42fb-4f4a-ba68-1686386b017e-openstack-config\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095602 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9kxz\" (UniqueName: \"kubernetes.io/projected/8ac368f3-42fb-4f4a-ba68-1686386b017e-kube-api-access-s9kxz\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.095620 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac368f3-42fb-4f4a-ba68-1686386b017e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.096279 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da936793-13b1-4815-a1ec-4d5d609ca5e3-logs" (OuterVolumeSpecName: "logs") pod "da936793-13b1-4815-a1ec-4d5d609ca5e3" (UID: "da936793-13b1-4815-a1ec-4d5d609ca5e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.102726 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "da936793-13b1-4815-a1ec-4d5d609ca5e3" (UID: "da936793-13b1-4815-a1ec-4d5d609ca5e3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.108493 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da936793-13b1-4815-a1ec-4d5d609ca5e3-kube-api-access-2m7fh" (OuterVolumeSpecName: "kube-api-access-2m7fh") pod "da936793-13b1-4815-a1ec-4d5d609ca5e3" (UID: "da936793-13b1-4815-a1ec-4d5d609ca5e3"). InnerVolumeSpecName "kube-api-access-2m7fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.149768 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da936793-13b1-4815-a1ec-4d5d609ca5e3" (UID: "da936793-13b1-4815-a1ec-4d5d609ca5e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.188901 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data" (OuterVolumeSpecName: "config-data") pod "da936793-13b1-4815-a1ec-4d5d609ca5e3" (UID: "da936793-13b1-4815-a1ec-4d5d609ca5e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.198625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ac368f3-42fb-4f4a-ba68-1686386b017e-openstack-config-secret\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.198871 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8ac368f3-42fb-4f4a-ba68-1686386b017e-openstack-config\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.199104 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9kxz\" (UniqueName: \"kubernetes.io/projected/8ac368f3-42fb-4f4a-ba68-1686386b017e-kube-api-access-s9kxz\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.199147 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac368f3-42fb-4f4a-ba68-1686386b017e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.199361 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.199406 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m7fh\" (UniqueName: \"kubernetes.io/projected/da936793-13b1-4815-a1ec-4d5d609ca5e3-kube-api-access-2m7fh\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.199429 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.199471 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/da936793-13b1-4815-a1ec-4d5d609ca5e3-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.199489 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da936793-13b1-4815-a1ec-4d5d609ca5e3-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.200667 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8ac368f3-42fb-4f4a-ba68-1686386b017e-openstack-config\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.206419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8ac368f3-42fb-4f4a-ba68-1686386b017e-openstack-config-secret\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.206978 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ac368f3-42fb-4f4a-ba68-1686386b017e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.218652 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9kxz\" (UniqueName: \"kubernetes.io/projected/8ac368f3-42fb-4f4a-ba68-1686386b017e-kube-api-access-s9kxz\") pod \"openstackclient\" (UID: \"8ac368f3-42fb-4f4a-ba68-1686386b017e\") " pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.374351 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.380473 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" event={"ID":"09eb7b3b-c5af-4625-8f1a-83766550711c","Type":"ContainerStarted","Data":"0bd4229260fb8fe792c73d8e69cf5442f931f736c0b59deda9ffc0fef5be3a78"} Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.380515 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" event={"ID":"09eb7b3b-c5af-4625-8f1a-83766550711c","Type":"ContainerStarted","Data":"f60997be790e511a986a9ab57afe7a3d5ed6fd416e480154b493f4664ba9afab"} Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.395585 4926 generic.go:334] "Generic (PLEG): container finished" podID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerID="7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8" exitCode=137 Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.395642 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.395641 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" event={"ID":"da936793-13b1-4815-a1ec-4d5d609ca5e3","Type":"ContainerDied","Data":"7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8"} Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.396369 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-94956785d-mtl2w" event={"ID":"da936793-13b1-4815-a1ec-4d5d609ca5e3","Type":"ContainerDied","Data":"01f48ef206a2d7f242f74c69621856e49171cd0d33097cfbec64d9c1935505bd"} Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.396403 4926 scope.go:117] "RemoveContainer" containerID="7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.552426 4926 scope.go:117] "RemoveContainer" containerID="38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.564746 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-94956785d-mtl2w"] Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.573661 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-94956785d-mtl2w"] Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.618074 4926 scope.go:117] "RemoveContainer" containerID="7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8" Mar 12 18:24:02 crc kubenswrapper[4926]: E0312 18:24:02.623502 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8\": container with ID starting with 7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8 not found: ID does not exist" containerID="7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.623557 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8"} err="failed to get container status \"7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8\": rpc error: code = NotFound desc = could not find container \"7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8\": container with ID starting with 7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8 not found: ID does not exist" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.623600 4926 scope.go:117] "RemoveContainer" containerID="38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2" Mar 12 18:24:02 crc kubenswrapper[4926]: E0312 18:24:02.624109 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2\": container with ID starting with 38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2 not found: ID does not exist" containerID="38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.624154 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2"} err="failed to get container status \"38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2\": rpc error: code = NotFound desc = could not find container \"38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2\": container with ID starting with 38486c735a1623333651d8064ff2479a19bcc2c864a0885d812a7528e1e135d2 not found: ID does not exist" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.861564 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.861863 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-central-agent" containerID="cri-o://a367820b72291dce3847ddbcc2b3a9f8493b17ddb9b80ab37f0c0b123bd5c72e" gracePeriod=30 Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.862507 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="proxy-httpd" containerID="cri-o://27c0e5b7f4011faa0c2b0ee4f54cfe990c1a5e30f0bbe162914f8380a857f6ef" gracePeriod=30 Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.862649 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-notification-agent" containerID="cri-o://a9eba08d9bcde00929ca468e28634454c84f19ace5c5074a2b2f0456e3403fef" gracePeriod=30 Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.862663 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="sg-core" containerID="cri-o://8a7449dd308bb99ab10627007eaa35755447eacebb3f5175eab4f716527295a3" gracePeriod=30 Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.873576 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.173:3000/\": EOF" Mar 12 18:24:02 crc kubenswrapper[4926]: I0312 18:24:02.913027 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.218089 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.406469 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8ac368f3-42fb-4f4a-ba68-1686386b017e","Type":"ContainerStarted","Data":"858d0c3a8121a26a1a23cb8cc32005f95c2ce3b6c8d15c7165700931ba74ca6f"} Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.409916 4926 generic.go:334] "Generic (PLEG): container finished" podID="d0a93b50-2038-4bf8-8c5f-bc77148d55f8" containerID="e0b6e5ac15bb107790c6fb8dd27667c6693088633c203e8fcc884a1e78a79525" exitCode=0 Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.409976 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555664-xqlj7" event={"ID":"d0a93b50-2038-4bf8-8c5f-bc77148d55f8","Type":"ContainerDied","Data":"e0b6e5ac15bb107790c6fb8dd27667c6693088633c203e8fcc884a1e78a79525"} Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.421418 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" event={"ID":"09eb7b3b-c5af-4625-8f1a-83766550711c","Type":"ContainerStarted","Data":"cb5574985a441766effb628172e6b7ca47066f959ac3872ce7b092cf6411844d"} Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.422714 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.422755 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447177 4926 generic.go:334] "Generic (PLEG): container finished" podID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerID="27c0e5b7f4011faa0c2b0ee4f54cfe990c1a5e30f0bbe162914f8380a857f6ef" exitCode=0 Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447212 4926 generic.go:334] "Generic (PLEG): container finished" podID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerID="8a7449dd308bb99ab10627007eaa35755447eacebb3f5175eab4f716527295a3" exitCode=2 Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447221 4926 generic.go:334] "Generic (PLEG): container finished" podID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerID="a9eba08d9bcde00929ca468e28634454c84f19ace5c5074a2b2f0456e3403fef" exitCode=0 Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447227 4926 generic.go:334] "Generic (PLEG): container finished" podID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerID="a367820b72291dce3847ddbcc2b3a9f8493b17ddb9b80ab37f0c0b123bd5c72e" exitCode=0 Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447286 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerDied","Data":"27c0e5b7f4011faa0c2b0ee4f54cfe990c1a5e30f0bbe162914f8380a857f6ef"} Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447326 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerDied","Data":"8a7449dd308bb99ab10627007eaa35755447eacebb3f5175eab4f716527295a3"} Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447336 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerDied","Data":"a9eba08d9bcde00929ca468e28634454c84f19ace5c5074a2b2f0456e3403fef"} Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.447345 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerDied","Data":"a367820b72291dce3847ddbcc2b3a9f8493b17ddb9b80ab37f0c0b123bd5c72e"} Mar 12 18:24:03 crc kubenswrapper[4926]: I0312 18:24:03.459286 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" podStartSLOduration=3.459256735 podStartE2EDuration="3.459256735s" podCreationTimestamp="2026-03-12 18:24:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:24:03.44350779 +0000 UTC m=+1283.812134123" watchObservedRunningTime="2026-03-12 18:24:03.459256735 +0000 UTC m=+1283.827883068" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.010915 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.182376 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-scripts\") pod \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.183067 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-combined-ca-bundle\") pod \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.183151 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bmcf\" (UniqueName: \"kubernetes.io/projected/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-kube-api-access-8bmcf\") pod \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.183720 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-log-httpd\") pod \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.183775 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-config-data\") pod \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.183840 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-sg-core-conf-yaml\") pod \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.183993 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-run-httpd\") pod \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\" (UID: \"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d\") " Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.184093 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" (UID: "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.185048 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.185282 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" (UID: "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.190340 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-kube-api-access-8bmcf" (OuterVolumeSpecName: "kube-api-access-8bmcf") pod "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" (UID: "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d"). InnerVolumeSpecName "kube-api-access-8bmcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.213709 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-scripts" (OuterVolumeSpecName: "scripts") pod "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" (UID: "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.224534 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" (UID: "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.287868 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.287897 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bmcf\" (UniqueName: \"kubernetes.io/projected/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-kube-api-access-8bmcf\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.287907 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.287921 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.288004 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" (UID: "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.312550 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-config-data" (OuterVolumeSpecName: "config-data") pod "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" (UID: "a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.389318 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.389369 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.467267 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d","Type":"ContainerDied","Data":"f1898b8c34b4b06f2e10cb4f43fffdb55131049338d63dbe39254ab53b9a2db0"} Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.467384 4926 scope.go:117] "RemoveContainer" containerID="27c0e5b7f4011faa0c2b0ee4f54cfe990c1a5e30f0bbe162914f8380a857f6ef" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.467759 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.499150 4926 scope.go:117] "RemoveContainer" containerID="8a7449dd308bb99ab10627007eaa35755447eacebb3f5175eab4f716527295a3" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.558072 4926 scope.go:117] "RemoveContainer" containerID="a9eba08d9bcde00929ca468e28634454c84f19ace5c5074a2b2f0456e3403fef" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.603543 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" path="/var/lib/kubelet/pods/da936793-13b1-4815-a1ec-4d5d609ca5e3/volumes" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.604287 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.604317 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.604336 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:04 crc kubenswrapper[4926]: E0312 18:24:04.605082 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605105 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener" Mar 12 18:24:04 crc kubenswrapper[4926]: E0312 18:24:04.605151 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-notification-agent" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605179 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-notification-agent" Mar 12 18:24:04 crc kubenswrapper[4926]: E0312 18:24:04.605205 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-central-agent" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605211 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-central-agent" Mar 12 18:24:04 crc kubenswrapper[4926]: E0312 18:24:04.605236 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener-log" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605243 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener-log" Mar 12 18:24:04 crc kubenswrapper[4926]: E0312 18:24:04.605257 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="proxy-httpd" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605266 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="proxy-httpd" Mar 12 18:24:04 crc kubenswrapper[4926]: E0312 18:24:04.605287 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="sg-core" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605293 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="sg-core" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605641 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605670 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-central-agent" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605696 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="proxy-httpd" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605711 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="sg-core" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605723 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" containerName="ceilometer-notification-agent" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.605738 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="da936793-13b1-4815-a1ec-4d5d609ca5e3" containerName="barbican-keystone-listener-log" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.609674 4926 scope.go:117] "RemoveContainer" containerID="a367820b72291dce3847ddbcc2b3a9f8493b17ddb9b80ab37f0c0b123bd5c72e" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.635897 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.636081 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.638879 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.639159 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.715107 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.715526 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-run-httpd\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.715629 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-log-httpd\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.715650 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-config-data\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.715704 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-scripts\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.715801 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.715828 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb8t5\" (UniqueName: \"kubernetes.io/projected/785370ab-aa78-4362-adf4-fedc8a0aedf9-kube-api-access-lb8t5\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.817300 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-scripts\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.817429 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.817492 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb8t5\" (UniqueName: \"kubernetes.io/projected/785370ab-aa78-4362-adf4-fedc8a0aedf9-kube-api-access-lb8t5\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.817540 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.817579 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-run-httpd\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.817736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-config-data\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.817760 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-log-httpd\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.818310 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-log-httpd\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.821080 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-run-httpd\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.826381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.832559 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.832715 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-config-data\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.834177 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-scripts\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.842213 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb8t5\" (UniqueName: \"kubernetes.io/projected/785370ab-aa78-4362-adf4-fedc8a0aedf9-kube-api-access-lb8t5\") pod \"ceilometer-0\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " pod="openstack/ceilometer-0" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.917117 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555664-xqlj7" Mar 12 18:24:04 crc kubenswrapper[4926]: I0312 18:24:04.971599 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.020716 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gznnp\" (UniqueName: \"kubernetes.io/projected/d0a93b50-2038-4bf8-8c5f-bc77148d55f8-kube-api-access-gznnp\") pod \"d0a93b50-2038-4bf8-8c5f-bc77148d55f8\" (UID: \"d0a93b50-2038-4bf8-8c5f-bc77148d55f8\") " Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.023834 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0a93b50-2038-4bf8-8c5f-bc77148d55f8-kube-api-access-gznnp" (OuterVolumeSpecName: "kube-api-access-gznnp") pod "d0a93b50-2038-4bf8-8c5f-bc77148d55f8" (UID: "d0a93b50-2038-4bf8-8c5f-bc77148d55f8"). InnerVolumeSpecName "kube-api-access-gznnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.123546 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gznnp\" (UniqueName: \"kubernetes.io/projected/d0a93b50-2038-4bf8-8c5f-bc77148d55f8-kube-api-access-gznnp\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.465883 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.482260 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555664-xqlj7" Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.482256 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555664-xqlj7" event={"ID":"d0a93b50-2038-4bf8-8c5f-bc77148d55f8","Type":"ContainerDied","Data":"b73f746aeaacae651cc800c031298955a2feeae1651f82a533f65b94b8415760"} Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.482336 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b73f746aeaacae651cc800c031298955a2feeae1651f82a533f65b94b8415760" Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.966569 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 12 18:24:05 crc kubenswrapper[4926]: I0312 18:24:05.993383 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555658-gfgnc"] Mar 12 18:24:06 crc kubenswrapper[4926]: I0312 18:24:06.008341 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555658-gfgnc"] Mar 12 18:24:06 crc kubenswrapper[4926]: I0312 18:24:06.506580 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b727d6f-6c9d-44ad-8594-accc9d2c4ed6" path="/var/lib/kubelet/pods/4b727d6f-6c9d-44ad-8594-accc9d2c4ed6/volumes" Mar 12 18:24:06 crc kubenswrapper[4926]: I0312 18:24:06.510368 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d" path="/var/lib/kubelet/pods/a64c41ad-a2b0-46ae-8d4c-fa0b01b8649d/volumes" Mar 12 18:24:06 crc kubenswrapper[4926]: I0312 18:24:06.512240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerStarted","Data":"122b7926eeb56b52d198ea8fb2b2bcfa65e9df33208dc1088b72c07833d13975"} Mar 12 18:24:06 crc kubenswrapper[4926]: I0312 18:24:06.512288 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerStarted","Data":"b830a701f318d482f59209b1a64738cbe2648c8287a45c52a5d2852a56c3420e"} Mar 12 18:24:07 crc kubenswrapper[4926]: I0312 18:24:07.503000 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerStarted","Data":"2edaaa27e784ed522859eae2d6ff079de37094800a6747ee19848273e81ce307"} Mar 12 18:24:08 crc kubenswrapper[4926]: I0312 18:24:08.515869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerStarted","Data":"fb25c14164a8c733a356a95d30484debd145ffe5ef07e6cd41f8828fbbe4674f"} Mar 12 18:24:11 crc kubenswrapper[4926]: I0312 18:24:11.257313 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:11 crc kubenswrapper[4926]: I0312 18:24:11.257651 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" Mar 12 18:24:12 crc kubenswrapper[4926]: I0312 18:24:12.129932 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:12 crc kubenswrapper[4926]: I0312 18:24:12.377292 4926 scope.go:117] "RemoveContainer" containerID="287cfd2e2e77a9b487765ae984d056f483b704272f97333956b53dc7f101b02d" Mar 12 18:24:13 crc kubenswrapper[4926]: I0312 18:24:13.190922 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7fc6496dbc-7qwrw" Mar 12 18:24:13 crc kubenswrapper[4926]: I0312 18:24:13.217719 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-89554fb64-s9c6q" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 12 18:24:13 crc kubenswrapper[4926]: I0312 18:24:13.298142 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-884f7b65b-tpkzl"] Mar 12 18:24:13 crc kubenswrapper[4926]: I0312 18:24:13.298346 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-884f7b65b-tpkzl" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-api" containerID="cri-o://e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84" gracePeriod=30 Mar 12 18:24:13 crc kubenswrapper[4926]: I0312 18:24:13.298930 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-884f7b65b-tpkzl" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-httpd" containerID="cri-o://d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529" gracePeriod=30 Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.573920 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"8ac368f3-42fb-4f4a-ba68-1686386b017e","Type":"ContainerStarted","Data":"5e31eedfcc536e64dcd654a034ae030854b30d0c69f093c2fd0e94322552616d"} Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.576921 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerStarted","Data":"a5363a011abc4afc5c22e60ffd3c571ee4c452c8a1940dab6a8d1fc67f9718cb"} Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.577096 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-central-agent" containerID="cri-o://122b7926eeb56b52d198ea8fb2b2bcfa65e9df33208dc1088b72c07833d13975" gracePeriod=30 Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.577162 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.577179 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="proxy-httpd" containerID="cri-o://a5363a011abc4afc5c22e60ffd3c571ee4c452c8a1940dab6a8d1fc67f9718cb" gracePeriod=30 Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.577222 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-notification-agent" containerID="cri-o://2edaaa27e784ed522859eae2d6ff079de37094800a6747ee19848273e81ce307" gracePeriod=30 Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.577282 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="sg-core" containerID="cri-o://fb25c14164a8c733a356a95d30484debd145ffe5ef07e6cd41f8828fbbe4674f" gracePeriod=30 Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.580159 4926 generic.go:334] "Generic (PLEG): container finished" podID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerID="d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529" exitCode=0 Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.580197 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-884f7b65b-tpkzl" event={"ID":"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea","Type":"ContainerDied","Data":"d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529"} Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.590886 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.803950591 podStartE2EDuration="12.590869652s" podCreationTimestamp="2026-03-12 18:24:02 +0000 UTC" firstStartedPulling="2026-03-12 18:24:02.893795322 +0000 UTC m=+1283.262421655" lastFinishedPulling="2026-03-12 18:24:13.680714383 +0000 UTC m=+1294.049340716" observedRunningTime="2026-03-12 18:24:14.590510601 +0000 UTC m=+1294.959136934" watchObservedRunningTime="2026-03-12 18:24:14.590869652 +0000 UTC m=+1294.959495985" Mar 12 18:24:14 crc kubenswrapper[4926]: I0312 18:24:14.620380 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.403011626 podStartE2EDuration="10.620355737s" podCreationTimestamp="2026-03-12 18:24:04 +0000 UTC" firstStartedPulling="2026-03-12 18:24:05.484472274 +0000 UTC m=+1285.853098607" lastFinishedPulling="2026-03-12 18:24:13.701816385 +0000 UTC m=+1294.070442718" observedRunningTime="2026-03-12 18:24:14.614416631 +0000 UTC m=+1294.983042984" watchObservedRunningTime="2026-03-12 18:24:14.620355737 +0000 UTC m=+1294.988982070" Mar 12 18:24:14 crc kubenswrapper[4926]: E0312 18:24:14.716241 4926 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/c2e53ff075e298480760552cda2544f1e9b4139d48e822beda05f83234508b59/diff" to get inode usage: stat /var/lib/containers/storage/overlay/c2e53ff075e298480760552cda2544f1e9b4139d48e822beda05f83234508b59/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_5fdfffa4-937c-4167-8545-d34f2007fbc9/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_5fdfffa4-937c-4167-8545-d34f2007fbc9/ceilometer-notification-agent/0.log: no such file or directory Mar 12 18:24:15 crc kubenswrapper[4926]: E0312 18:24:15.338133 4926 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/eb3ee65486eee702d32220ece4785ba2ffa0d61d40fd91e9ccbf68ec12ce3c76/diff" to get inode usage: stat /var/lib/containers/storage/overlay/eb3ee65486eee702d32220ece4785ba2ffa0d61d40fd91e9ccbf68ec12ce3c76/diff: no such file or directory, extraDiskErr: Mar 12 18:24:15 crc kubenswrapper[4926]: E0312 18:24:15.498654 4926 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/3848119c075ad08b2848fa09f9b39831f837a8bb14ca7b0747e0d453e8757f2e/diff" to get inode usage: stat /var/lib/containers/storage/overlay/3848119c075ad08b2848fa09f9b39831f837a8bb14ca7b0747e0d453e8757f2e/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-65c5c86775-mct68_2beed02e-2edf-4b52-8ea6-ae2dae7502d8/neutron-api/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-65c5c86775-mct68_2beed02e-2edf-4b52-8ea6-ae2dae7502d8/neutron-api/0.log: no such file or directory Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597578 4926 generic.go:334] "Generic (PLEG): container finished" podID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerID="a5363a011abc4afc5c22e60ffd3c571ee4c452c8a1940dab6a8d1fc67f9718cb" exitCode=0 Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597624 4926 generic.go:334] "Generic (PLEG): container finished" podID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerID="fb25c14164a8c733a356a95d30484debd145ffe5ef07e6cd41f8828fbbe4674f" exitCode=2 Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597635 4926 generic.go:334] "Generic (PLEG): container finished" podID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerID="2edaaa27e784ed522859eae2d6ff079de37094800a6747ee19848273e81ce307" exitCode=0 Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597634 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerDied","Data":"a5363a011abc4afc5c22e60ffd3c571ee4c452c8a1940dab6a8d1fc67f9718cb"} Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597685 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerDied","Data":"fb25c14164a8c733a356a95d30484debd145ffe5ef07e6cd41f8828fbbe4674f"} Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597700 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerDied","Data":"2edaaa27e784ed522859eae2d6ff079de37094800a6747ee19848273e81ce307"} Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597713 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerDied","Data":"122b7926eeb56b52d198ea8fb2b2bcfa65e9df33208dc1088b72c07833d13975"} Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.597649 4926 generic.go:334] "Generic (PLEG): container finished" podID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerID="122b7926eeb56b52d198ea8fb2b2bcfa65e9df33208dc1088b72c07833d13975" exitCode=0 Mar 12 18:24:15 crc kubenswrapper[4926]: E0312 18:24:15.777537 4926 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/83a98e7d89ad57b3c990c9d485f73210c416f72d87846ea5492e22a752654c27/diff" to get inode usage: stat /var/lib/containers/storage/overlay/83a98e7d89ad57b3c990c9d485f73210c416f72d87846ea5492e22a752654c27/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-65c5c86775-mct68_2beed02e-2edf-4b52-8ea6-ae2dae7502d8/neutron-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-65c5c86775-mct68_2beed02e-2edf-4b52-8ea6-ae2dae7502d8/neutron-httpd/0.log: no such file or directory Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.786130 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.849405 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-config-data\") pod \"785370ab-aa78-4362-adf4-fedc8a0aedf9\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.849587 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-combined-ca-bundle\") pod \"785370ab-aa78-4362-adf4-fedc8a0aedf9\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.849609 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-log-httpd\") pod \"785370ab-aa78-4362-adf4-fedc8a0aedf9\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.849696 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-scripts\") pod \"785370ab-aa78-4362-adf4-fedc8a0aedf9\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.849727 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb8t5\" (UniqueName: \"kubernetes.io/projected/785370ab-aa78-4362-adf4-fedc8a0aedf9-kube-api-access-lb8t5\") pod \"785370ab-aa78-4362-adf4-fedc8a0aedf9\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.849762 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-sg-core-conf-yaml\") pod \"785370ab-aa78-4362-adf4-fedc8a0aedf9\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.849807 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-run-httpd\") pod \"785370ab-aa78-4362-adf4-fedc8a0aedf9\" (UID: \"785370ab-aa78-4362-adf4-fedc8a0aedf9\") " Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.850747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "785370ab-aa78-4362-adf4-fedc8a0aedf9" (UID: "785370ab-aa78-4362-adf4-fedc8a0aedf9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.851413 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "785370ab-aa78-4362-adf4-fedc8a0aedf9" (UID: "785370ab-aa78-4362-adf4-fedc8a0aedf9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.854761 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-scripts" (OuterVolumeSpecName: "scripts") pod "785370ab-aa78-4362-adf4-fedc8a0aedf9" (UID: "785370ab-aa78-4362-adf4-fedc8a0aedf9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.869783 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/785370ab-aa78-4362-adf4-fedc8a0aedf9-kube-api-access-lb8t5" (OuterVolumeSpecName: "kube-api-access-lb8t5") pod "785370ab-aa78-4362-adf4-fedc8a0aedf9" (UID: "785370ab-aa78-4362-adf4-fedc8a0aedf9"). InnerVolumeSpecName "kube-api-access-lb8t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.901152 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "785370ab-aa78-4362-adf4-fedc8a0aedf9" (UID: "785370ab-aa78-4362-adf4-fedc8a0aedf9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.951726 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.951767 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb8t5\" (UniqueName: \"kubernetes.io/projected/785370ab-aa78-4362-adf4-fedc8a0aedf9-kube-api-access-lb8t5\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.951781 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.951792 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.951804 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/785370ab-aa78-4362-adf4-fedc8a0aedf9-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.955794 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "785370ab-aa78-4362-adf4-fedc8a0aedf9" (UID: "785370ab-aa78-4362-adf4-fedc8a0aedf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:15 crc kubenswrapper[4926]: I0312 18:24:15.975604 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-config-data" (OuterVolumeSpecName: "config-data") pod "785370ab-aa78-4362-adf4-fedc8a0aedf9" (UID: "785370ab-aa78-4362-adf4-fedc8a0aedf9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.053359 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.053390 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/785370ab-aa78-4362-adf4-fedc8a0aedf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.607107 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"785370ab-aa78-4362-adf4-fedc8a0aedf9","Type":"ContainerDied","Data":"b830a701f318d482f59209b1a64738cbe2648c8287a45c52a5d2852a56c3420e"} Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.607380 4926 scope.go:117] "RemoveContainer" containerID="a5363a011abc4afc5c22e60ffd3c571ee4c452c8a1940dab6a8d1fc67f9718cb" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.607302 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.633129 4926 scope.go:117] "RemoveContainer" containerID="fb25c14164a8c733a356a95d30484debd145ffe5ef07e6cd41f8828fbbe4674f" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.639487 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.648242 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.660760 4926 scope.go:117] "RemoveContainer" containerID="2edaaa27e784ed522859eae2d6ff079de37094800a6747ee19848273e81ce307" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.668961 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:16 crc kubenswrapper[4926]: E0312 18:24:16.669352 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0a93b50-2038-4bf8-8c5f-bc77148d55f8" containerName="oc" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669372 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0a93b50-2038-4bf8-8c5f-bc77148d55f8" containerName="oc" Mar 12 18:24:16 crc kubenswrapper[4926]: E0312 18:24:16.669386 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="proxy-httpd" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669392 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="proxy-httpd" Mar 12 18:24:16 crc kubenswrapper[4926]: E0312 18:24:16.669402 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-central-agent" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669408 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-central-agent" Mar 12 18:24:16 crc kubenswrapper[4926]: E0312 18:24:16.669419 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-notification-agent" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669425 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-notification-agent" Mar 12 18:24:16 crc kubenswrapper[4926]: E0312 18:24:16.669450 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="sg-core" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669457 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="sg-core" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669624 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0a93b50-2038-4bf8-8c5f-bc77148d55f8" containerName="oc" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669637 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-central-agent" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669649 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="ceilometer-notification-agent" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669662 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="proxy-httpd" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.669673 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" containerName="sg-core" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.671379 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.674311 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.674513 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.690765 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.743383 4926 scope.go:117] "RemoveContainer" containerID="122b7926eeb56b52d198ea8fb2b2bcfa65e9df33208dc1088b72c07833d13975" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.768165 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.768210 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b9v4\" (UniqueName: \"kubernetes.io/projected/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-kube-api-access-7b9v4\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.768275 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-run-httpd\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.768310 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-scripts\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.768333 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-log-httpd\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.768373 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.768448 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-config-data\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.871628 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-run-httpd\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.871710 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-scripts\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.871738 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-log-httpd\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.871782 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.871819 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-config-data\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.871894 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.871927 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b9v4\" (UniqueName: \"kubernetes.io/projected/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-kube-api-access-7b9v4\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.872583 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-log-httpd\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.873933 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-run-httpd\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.878928 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-scripts\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.879157 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.880020 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-config-data\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.880057 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:16 crc kubenswrapper[4926]: I0312 18:24:16.892419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b9v4\" (UniqueName: \"kubernetes.io/projected/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-kube-api-access-7b9v4\") pod \"ceilometer-0\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " pod="openstack/ceilometer-0" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.098004 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.309665 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.380137 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vz6v\" (UniqueName: \"kubernetes.io/projected/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-kube-api-access-7vz6v\") pod \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.380194 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-httpd-config\") pod \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.380247 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-combined-ca-bundle\") pod \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.380291 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-config\") pod \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.380319 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-ovndb-tls-certs\") pod \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\" (UID: \"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea\") " Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.386871 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" (UID: "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.391668 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-kube-api-access-7vz6v" (OuterVolumeSpecName: "kube-api-access-7vz6v") pod "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" (UID: "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea"). InnerVolumeSpecName "kube-api-access-7vz6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.437255 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" (UID: "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.444841 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-config" (OuterVolumeSpecName: "config") pod "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" (UID: "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.451748 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" (UID: "5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.482190 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vz6v\" (UniqueName: \"kubernetes.io/projected/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-kube-api-access-7vz6v\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.482224 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.482234 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.482244 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.482252 4926 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.588130 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.616415 4926 generic.go:334] "Generic (PLEG): container finished" podID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerID="e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84" exitCode=0 Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.616491 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-884f7b65b-tpkzl" event={"ID":"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea","Type":"ContainerDied","Data":"e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84"} Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.616525 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-884f7b65b-tpkzl" event={"ID":"5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea","Type":"ContainerDied","Data":"4598f84bb743a934c385d17beee4641ea28d3b613e4146bda41f42d7c8f987e3"} Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.616547 4926 scope.go:117] "RemoveContainer" containerID="d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.616647 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-884f7b65b-tpkzl" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.631311 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerStarted","Data":"7996e6fe65ac453b2bf7a0036550739c5607aa7bde6261bcc4407d93c97a4ca7"} Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.655501 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-884f7b65b-tpkzl"] Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.657304 4926 scope.go:117] "RemoveContainer" containerID="e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.664503 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-884f7b65b-tpkzl"] Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.677814 4926 scope.go:117] "RemoveContainer" containerID="d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529" Mar 12 18:24:17 crc kubenswrapper[4926]: E0312 18:24:17.678184 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529\": container with ID starting with d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529 not found: ID does not exist" containerID="d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.678237 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529"} err="failed to get container status \"d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529\": rpc error: code = NotFound desc = could not find container \"d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529\": container with ID starting with d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529 not found: ID does not exist" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.678268 4926 scope.go:117] "RemoveContainer" containerID="e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84" Mar 12 18:24:17 crc kubenswrapper[4926]: E0312 18:24:17.678602 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84\": container with ID starting with e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84 not found: ID does not exist" containerID="e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84" Mar 12 18:24:17 crc kubenswrapper[4926]: I0312 18:24:17.678633 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84"} err="failed to get container status \"e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84\": rpc error: code = NotFound desc = could not find container \"e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84\": container with ID starting with e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84 not found: ID does not exist" Mar 12 18:24:18 crc kubenswrapper[4926]: E0312 18:24:18.421231 4926 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: , extraDiskErr: could not stat "/var/log/pods/openstack_cinder-api-0_78cc2e75-39e6-4148-87a5-022cc3690da8/cinder-api-log/0.log" to get inode usage: stat /var/log/pods/openstack_cinder-api-0_78cc2e75-39e6-4148-87a5-022cc3690da8/cinder-api-log/0.log: no such file or directory Mar 12 18:24:18 crc kubenswrapper[4926]: W0312 18:24:18.421420 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a13f5e0_72f6_4c47_a5ea_349c2d618d8c.slice/crio-5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc.scope WatchSource:0}: Error finding container 5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc: Status 404 returned error can't find the container with id 5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc Mar 12 18:24:18 crc kubenswrapper[4926]: E0312 18:24:18.432486 4926 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f1f2a42_878e_46c0_bd66_4927a4689299.slice/crio-28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc: Error finding container 28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc: Status 404 returned error can't find the container with id 28a435d769e66430d190ede24f5f603701f099bd6688244c585cc8bf6e7750fc Mar 12 18:24:18 crc kubenswrapper[4926]: E0312 18:24:18.432517 4926 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: , extraDiskErr: could not stat "/var/log/pods/openstack_cinder-api-0_78cc2e75-39e6-4148-87a5-022cc3690da8/cinder-api/0.log" to get inode usage: stat /var/log/pods/openstack_cinder-api-0_78cc2e75-39e6-4148-87a5-022cc3690da8/cinder-api/0.log: no such file or directory Mar 12 18:24:18 crc kubenswrapper[4926]: W0312 18:24:18.434236 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a13f5e0_72f6_4c47_a5ea_349c2d618d8c.slice/crio-433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74.scope WatchSource:0}: Error finding container 433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74: Status 404 returned error can't find the container with id 433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74 Mar 12 18:24:18 crc kubenswrapper[4926]: W0312 18:24:18.434930 4926 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda64c41ad_a2b0_46ae_8d4c_fa0b01b8649d.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda64c41ad_a2b0_46ae_8d4c_fa0b01b8649d.slice: no such file or directory Mar 12 18:24:18 crc kubenswrapper[4926]: W0312 18:24:18.478890 4926 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0a93b50_2038_4bf8_8c5f_bc77148d55f8.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0a93b50_2038_4bf8_8c5f_bc77148d55f8.slice: no such file or directory Mar 12 18:24:18 crc kubenswrapper[4926]: W0312 18:24:18.499830 4926 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod785370ab_aa78_4362_adf4_fedc8a0aedf9.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod785370ab_aa78_4362_adf4_fedc8a0aedf9.slice: no such file or directory Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.529006 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" path="/var/lib/kubelet/pods/5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea/volumes" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.530853 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="785370ab-aa78-4362-adf4-fedc8a0aedf9" path="/var/lib/kubelet/pods/785370ab-aa78-4362-adf4-fedc8a0aedf9/volumes" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.598326 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-55g54"] Mar 12 18:24:18 crc kubenswrapper[4926]: E0312 18:24:18.601963 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-httpd" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.601998 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-httpd" Mar 12 18:24:18 crc kubenswrapper[4926]: E0312 18:24:18.602036 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-api" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.602062 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-api" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.602274 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-api" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.602287 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc8f91a-7f34-4957-b3e4-e5d28b2f43ea" containerName="neutron-httpd" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.609059 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.619985 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-55g54"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.656045 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerStarted","Data":"70089f60f5708dc42a17c61910ff76a140c92fff334f5b497df2b30ef8db681d"} Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.678201 4926 generic.go:334] "Generic (PLEG): container finished" podID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerID="a41e698428edf8a4dccc211b474a012677761bd1a52d186e37df8c3d1445ee62" exitCode=137 Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.678307 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-89554fb64-s9c6q" event={"ID":"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572","Type":"ContainerDied","Data":"a41e698428edf8a4dccc211b474a012677761bd1a52d186e37df8c3d1445ee62"} Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.704089 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-t4krn"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.710249 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.713111 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb8f27f-336b-4acc-8165-0c53c1643084-operator-scripts\") pod \"nova-api-db-create-55g54\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.713209 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm9pf\" (UniqueName: \"kubernetes.io/projected/4cb8f27f-336b-4acc-8165-0c53c1643084-kube-api-access-jm9pf\") pod \"nova-api-db-create-55g54\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.734706 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-382f-account-create-update-7sn9z"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.736975 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.739244 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.748621 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-t4krn"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.762574 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-382f-account-create-update-7sn9z"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.816638 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb8f27f-336b-4acc-8165-0c53c1643084-operator-scripts\") pod \"nova-api-db-create-55g54\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.816699 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjfpj\" (UniqueName: \"kubernetes.io/projected/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-kube-api-access-hjfpj\") pod \"nova-api-382f-account-create-update-7sn9z\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.816759 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbjlt\" (UniqueName: \"kubernetes.io/projected/15db9185-0441-431d-98db-5701f9b244be-kube-api-access-gbjlt\") pod \"nova-cell0-db-create-t4krn\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.816792 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm9pf\" (UniqueName: \"kubernetes.io/projected/4cb8f27f-336b-4acc-8165-0c53c1643084-kube-api-access-jm9pf\") pod \"nova-api-db-create-55g54\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.816825 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-operator-scripts\") pod \"nova-api-382f-account-create-update-7sn9z\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.816862 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15db9185-0441-431d-98db-5701f9b244be-operator-scripts\") pod \"nova-cell0-db-create-t4krn\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.817824 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb8f27f-336b-4acc-8165-0c53c1643084-operator-scripts\") pod \"nova-api-db-create-55g54\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.845683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm9pf\" (UniqueName: \"kubernetes.io/projected/4cb8f27f-336b-4acc-8165-0c53c1643084-kube-api-access-jm9pf\") pod \"nova-api-db-create-55g54\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.898511 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-cctxg"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.900051 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.919184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-operator-scripts\") pod \"nova-cell1-db-create-cctxg\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.919293 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjfpj\" (UniqueName: \"kubernetes.io/projected/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-kube-api-access-hjfpj\") pod \"nova-api-382f-account-create-update-7sn9z\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.919413 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbjlt\" (UniqueName: \"kubernetes.io/projected/15db9185-0441-431d-98db-5701f9b244be-kube-api-access-gbjlt\") pod \"nova-cell0-db-create-t4krn\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.919494 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-operator-scripts\") pod \"nova-api-382f-account-create-update-7sn9z\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.919577 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15db9185-0441-431d-98db-5701f9b244be-operator-scripts\") pod \"nova-cell0-db-create-t4krn\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.919665 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxrpx\" (UniqueName: \"kubernetes.io/projected/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-kube-api-access-mxrpx\") pod \"nova-cell1-db-create-cctxg\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.920578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15db9185-0441-431d-98db-5701f9b244be-operator-scripts\") pod \"nova-cell0-db-create-t4krn\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.920855 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-operator-scripts\") pod \"nova-api-382f-account-create-update-7sn9z\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.923618 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f45b-account-create-update-rt9j9"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.925105 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.926739 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.941819 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-cctxg"] Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.942150 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjfpj\" (UniqueName: \"kubernetes.io/projected/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-kube-api-access-hjfpj\") pod \"nova-api-382f-account-create-update-7sn9z\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.942629 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.944384 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbjlt\" (UniqueName: \"kubernetes.io/projected/15db9185-0441-431d-98db-5701f9b244be-kube-api-access-gbjlt\") pod \"nova-cell0-db-create-t4krn\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:18 crc kubenswrapper[4926]: I0312 18:24:18.953311 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f45b-account-create-update-rt9j9"] Mar 12 18:24:18 crc kubenswrapper[4926]: E0312 18:24:18.988133 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2beed02e_2edf_4b52_8ea6_ae2dae7502d8.slice/crio-94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc8f91a_7f34_4957_b3e4_e5d28b2f43ea.slice/crio-d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda936793_13b1_4815_a1ec_4d5d609ca5e3.slice/crio-7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7d07aa_8c5e_49f3_8d85_4c5e9569c572.slice/crio-2388379aefe6f54987e9387c0ab60e55776d5a81c5e2c5b0f21608b48b6e8fa5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7d07aa_8c5e_49f3_8d85_4c5e9569c572.slice/crio-conmon-a41e698428edf8a4dccc211b474a012677761bd1a52d186e37df8c3d1445ee62.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7d07aa_8c5e_49f3_8d85_4c5e9569c572.slice/crio-conmon-2388379aefe6f54987e9387c0ab60e55776d5a81c5e2c5b0f21608b48b6e8fa5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda936793_13b1_4815_a1ec_4d5d609ca5e3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08abfe56_0e5c_4634_9a1a_488e2bbb587d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda936793_13b1_4815_a1ec_4d5d609ca5e3.slice/crio-01f48ef206a2d7f242f74c69621856e49171cd0d33097cfbec64d9c1935505bd\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2beed02e_2edf_4b52_8ea6_ae2dae7502d8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a13f5e0_72f6_4c47_a5ea_349c2d618d8c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc8f91a_7f34_4957_b3e4_e5d28b2f43ea.slice/crio-e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc7d07aa_8c5e_49f3_8d85_4c5e9569c572.slice/crio-a41e698428edf8a4dccc211b474a012677761bd1a52d186e37df8c3d1445ee62.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc8f91a_7f34_4957_b3e4_e5d28b2f43ea.slice/crio-conmon-e8a41592cfee0a29605ce4445cc98b991801ec37b06cc31b26495c7cc367ec84.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc8f91a_7f34_4957_b3e4_e5d28b2f43ea.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc8f91a_7f34_4957_b3e4_e5d28b2f43ea.slice/crio-4598f84bb743a934c385d17beee4641ea28d3b613e4146bda41f42d7c8f987e3\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dc8f91a_7f34_4957_b3e4_e5d28b2f43ea.slice/crio-conmon-d0514254b366a8c0de092253a4eb8c31e2072ea4a0ebc8efb64f738ddebe6529.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda936793_13b1_4815_a1ec_4d5d609ca5e3.slice/crio-conmon-7d7476ed20c94347bce99f3f0ec729c3292d3ed1f2ba8fd6cc877698a95ad5b8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2beed02e_2edf_4b52_8ea6_ae2dae7502d8.slice/crio-conmon-94b08a85a6e907ae8377090a58dcd3e4401da2dbe17a5ae03f9e88d0b3d26f8e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a13f5e0_72f6_4c47_a5ea_349c2d618d8c.slice/crio-conmon-5a5c133b5ffc1b7921863c9e06663dd92d6d307bfb1967e1d619b5cabd0d99dc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08abfe56_0e5c_4634_9a1a_488e2bbb587d.slice/crio-conmon-3e365f4cb50835b1180ff7c19d7fc397df5fff675446efdb00f5237496ddec2b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08abfe56_0e5c_4634_9a1a_488e2bbb587d.slice/crio-07e4278fe40ae8695fb94d6a2d5bec78c86b69a3255ccbd19aa22a7865e54876\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2beed02e_2edf_4b52_8ea6_ae2dae7502d8.slice/crio-7b5dc2f13fd79b24bbd53ad5c8577f6fcd9167f1b3fea4e0a8139bb248f187ac\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a13f5e0_72f6_4c47_a5ea_349c2d618d8c.slice/crio-conmon-433bb82141391ed92a9e550d2edd35bdc32cd216136772372e6f96c9b643ee74.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08abfe56_0e5c_4634_9a1a_488e2bbb587d.slice/crio-3e365f4cb50835b1180ff7c19d7fc397df5fff675446efdb00f5237496ddec2b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f1f2a42_878e_46c0_bd66_4927a4689299.slice\": RecentStats: unable to find data in memory cache]" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.015233 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.023817 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-operator-scripts\") pod \"nova-cell0-f45b-account-create-update-rt9j9\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.023909 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxrpx\" (UniqueName: \"kubernetes.io/projected/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-kube-api-access-mxrpx\") pod \"nova-cell1-db-create-cctxg\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.023974 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rll4\" (UniqueName: \"kubernetes.io/projected/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-kube-api-access-5rll4\") pod \"nova-cell0-f45b-account-create-update-rt9j9\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.024018 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-operator-scripts\") pod \"nova-cell1-db-create-cctxg\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.024729 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-operator-scripts\") pod \"nova-cell1-db-create-cctxg\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.026687 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.062818 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxrpx\" (UniqueName: \"kubernetes.io/projected/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-kube-api-access-mxrpx\") pod \"nova-cell1-db-create-cctxg\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.084366 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.112425 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-0630-account-create-update-knx9n"] Mar 12 18:24:19 crc kubenswrapper[4926]: E0312 18:24:19.112993 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon-log" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.113004 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon-log" Mar 12 18:24:19 crc kubenswrapper[4926]: E0312 18:24:19.113029 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.113035 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.113195 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.113205 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" containerName="horizon-log" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.113718 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.114907 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0630-account-create-update-knx9n"] Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.119546 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.125640 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-secret-key\") pod \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.125684 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-scripts\") pod \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.125705 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-logs\") pod \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.125864 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rll4\" (UniqueName: \"kubernetes.io/projected/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-kube-api-access-5rll4\") pod \"nova-cell0-f45b-account-create-update-rt9j9\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.125891 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29a2618b-1e0a-4246-b63c-582d2fe2f847-operator-scripts\") pod \"nova-cell1-0630-account-create-update-knx9n\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.125961 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8qwj\" (UniqueName: \"kubernetes.io/projected/29a2618b-1e0a-4246-b63c-582d2fe2f847-kube-api-access-p8qwj\") pod \"nova-cell1-0630-account-create-update-knx9n\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.126002 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-operator-scripts\") pod \"nova-cell0-f45b-account-create-update-rt9j9\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.126687 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-operator-scripts\") pod \"nova-cell0-f45b-account-create-update-rt9j9\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.127710 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-logs" (OuterVolumeSpecName: "logs") pod "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" (UID: "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.135811 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" (UID: "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.148927 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rll4\" (UniqueName: \"kubernetes.io/projected/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-kube-api-access-5rll4\") pod \"nova-cell0-f45b-account-create-update-rt9j9\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.195200 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-scripts" (OuterVolumeSpecName: "scripts") pod "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" (UID: "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227136 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p78br\" (UniqueName: \"kubernetes.io/projected/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-kube-api-access-p78br\") pod \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227185 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-tls-certs\") pod \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227224 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-config-data\") pod \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227260 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-combined-ca-bundle\") pod \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\" (UID: \"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572\") " Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227483 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29a2618b-1e0a-4246-b63c-582d2fe2f847-operator-scripts\") pod \"nova-cell1-0630-account-create-update-knx9n\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227549 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8qwj\" (UniqueName: \"kubernetes.io/projected/29a2618b-1e0a-4246-b63c-582d2fe2f847-kube-api-access-p8qwj\") pod \"nova-cell1-0630-account-create-update-knx9n\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227642 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227659 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.227669 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.228275 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29a2618b-1e0a-4246-b63c-582d2fe2f847-operator-scripts\") pod \"nova-cell1-0630-account-create-update-knx9n\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.233359 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-kube-api-access-p78br" (OuterVolumeSpecName: "kube-api-access-p78br") pod "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" (UID: "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572"). InnerVolumeSpecName "kube-api-access-p78br". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.243578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8qwj\" (UniqueName: \"kubernetes.io/projected/29a2618b-1e0a-4246-b63c-582d2fe2f847-kube-api-access-p8qwj\") pod \"nova-cell1-0630-account-create-update-knx9n\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.252915 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-config-data" (OuterVolumeSpecName: "config-data") pod "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" (UID: "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.294642 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" (UID: "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.317539 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" (UID: "dc7d07aa-8c5e-49f3-8d85-4c5e9569c572"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.328949 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.329004 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p78br\" (UniqueName: \"kubernetes.io/projected/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-kube-api-access-p78br\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.329042 4926 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.329055 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.329068 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.341639 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.435288 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.462745 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-382f-account-create-update-7sn9z"] Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.588708 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-55g54"] Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.660581 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-t4krn"] Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.757810 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-89554fb64-s9c6q" event={"ID":"dc7d07aa-8c5e-49f3-8d85-4c5e9569c572","Type":"ContainerDied","Data":"47a3a3a3dc751bc7ac0243d80b1ecc826d87bf09a948cad8eeaa84ccc31fd39c"} Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.757856 4926 scope.go:117] "RemoveContainer" containerID="2388379aefe6f54987e9387c0ab60e55776d5a81c5e2c5b0f21608b48b6e8fa5" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.757996 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-89554fb64-s9c6q" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.776735 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t4krn" event={"ID":"15db9185-0441-431d-98db-5701f9b244be","Type":"ContainerStarted","Data":"6c9449dc088a16c50ac498f087e95dd4074e327717458b2a2dca40a15417614a"} Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.793601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerStarted","Data":"309dce8d3efcb58229ff7be17feec841c39fb231db3ee8c9eaa9375400b8f496"} Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.800769 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-cctxg"] Mar 12 18:24:19 crc kubenswrapper[4926]: W0312 18:24:19.822808 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc4b3e47_ab9e_46c0_be37_9d2e94721be6.slice/crio-e0b512c7b0ec080f65f1af63e533a0195e053d720c63ac72bbcdd0918ebd9be0 WatchSource:0}: Error finding container e0b512c7b0ec080f65f1af63e533a0195e053d720c63ac72bbcdd0918ebd9be0: Status 404 returned error can't find the container with id e0b512c7b0ec080f65f1af63e533a0195e053d720c63ac72bbcdd0918ebd9be0 Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.822914 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-382f-account-create-update-7sn9z" event={"ID":"fe0c0d35-e463-4b5c-a15f-cdfc808c498c","Type":"ContainerStarted","Data":"b1d04c89d3d9fb6caf62bdb2b456380c91864ceab753f46a51bad242a32a7521"} Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.832537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-55g54" event={"ID":"4cb8f27f-336b-4acc-8165-0c53c1643084","Type":"ContainerStarted","Data":"c6f0363211c7d4e74e3ccbba5ec4240749e8ebe5592c2510b2fa60c1f08ec3c9"} Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.848552 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-382f-account-create-update-7sn9z" podStartSLOduration=1.848534211 podStartE2EDuration="1.848534211s" podCreationTimestamp="2026-03-12 18:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:24:19.836514084 +0000 UTC m=+1300.205140407" watchObservedRunningTime="2026-03-12 18:24:19.848534211 +0000 UTC m=+1300.217160544" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.853095 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-55g54" podStartSLOduration=1.853076104 podStartE2EDuration="1.853076104s" podCreationTimestamp="2026-03-12 18:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:24:19.849302775 +0000 UTC m=+1300.217929118" watchObservedRunningTime="2026-03-12 18:24:19.853076104 +0000 UTC m=+1300.221702427" Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.918524 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f45b-account-create-update-rt9j9"] Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.960248 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-89554fb64-s9c6q"] Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.975415 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-89554fb64-s9c6q"] Mar 12 18:24:19 crc kubenswrapper[4926]: I0312 18:24:19.993653 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-0630-account-create-update-knx9n"] Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.077657 4926 scope.go:117] "RemoveContainer" containerID="a41e698428edf8a4dccc211b474a012677761bd1a52d186e37df8c3d1445ee62" Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.507333 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc7d07aa-8c5e-49f3-8d85-4c5e9569c572" path="/var/lib/kubelet/pods/dc7d07aa-8c5e-49f3-8d85-4c5e9569c572/volumes" Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.852242 4926 generic.go:334] "Generic (PLEG): container finished" podID="29a2618b-1e0a-4246-b63c-582d2fe2f847" containerID="b0665b4cdb9e47cd07d74cad2d92a7eeb8aadf990846b2f50cc7f0a3a493d8dd" exitCode=0 Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.852747 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0630-account-create-update-knx9n" event={"ID":"29a2618b-1e0a-4246-b63c-582d2fe2f847","Type":"ContainerDied","Data":"b0665b4cdb9e47cd07d74cad2d92a7eeb8aadf990846b2f50cc7f0a3a493d8dd"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.852794 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0630-account-create-update-knx9n" event={"ID":"29a2618b-1e0a-4246-b63c-582d2fe2f847","Type":"ContainerStarted","Data":"ff34501c9af340574663fc5763d61f95a310be34a280df0b21b236386141ee47"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.855520 4926 generic.go:334] "Generic (PLEG): container finished" podID="dc4b3e47-ab9e-46c0-be37-9d2e94721be6" containerID="8476342456babac729c7086b287eb9b60e559eb560bedd1d78d42784e15f0ff2" exitCode=0 Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.855564 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cctxg" event={"ID":"dc4b3e47-ab9e-46c0-be37-9d2e94721be6","Type":"ContainerDied","Data":"8476342456babac729c7086b287eb9b60e559eb560bedd1d78d42784e15f0ff2"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.855584 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cctxg" event={"ID":"dc4b3e47-ab9e-46c0-be37-9d2e94721be6","Type":"ContainerStarted","Data":"e0b512c7b0ec080f65f1af63e533a0195e053d720c63ac72bbcdd0918ebd9be0"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.857314 4926 generic.go:334] "Generic (PLEG): container finished" podID="15db9185-0441-431d-98db-5701f9b244be" containerID="a9bf2ae52889d181443948b724481f92343c2befd9093bc8cc132900dface592" exitCode=0 Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.857349 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t4krn" event={"ID":"15db9185-0441-431d-98db-5701f9b244be","Type":"ContainerDied","Data":"a9bf2ae52889d181443948b724481f92343c2befd9093bc8cc132900dface592"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.859127 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerStarted","Data":"ef9ed90f72e8dfcc270f36331c95e42d9aa6057e9d8ae052ebb5501d32f7648b"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.864452 4926 generic.go:334] "Generic (PLEG): container finished" podID="fe0c0d35-e463-4b5c-a15f-cdfc808c498c" containerID="976b5bff6bf1beb72cdab8caedbb157ee653f67334ae07735bc5919349b4ae10" exitCode=0 Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.864497 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-382f-account-create-update-7sn9z" event={"ID":"fe0c0d35-e463-4b5c-a15f-cdfc808c498c","Type":"ContainerDied","Data":"976b5bff6bf1beb72cdab8caedbb157ee653f67334ae07735bc5919349b4ae10"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.866245 4926 generic.go:334] "Generic (PLEG): container finished" podID="b33ff126-1b0c-47f0-a4c1-18e1297fa81d" containerID="f80b93ed7cbf010c51c1bbf4a01b83fdc0dc54adf44c99408ef3f7476eea541b" exitCode=0 Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.866305 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" event={"ID":"b33ff126-1b0c-47f0-a4c1-18e1297fa81d","Type":"ContainerDied","Data":"f80b93ed7cbf010c51c1bbf4a01b83fdc0dc54adf44c99408ef3f7476eea541b"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.866319 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" event={"ID":"b33ff126-1b0c-47f0-a4c1-18e1297fa81d","Type":"ContainerStarted","Data":"198fc93efd6dbc2346260a021b2acdc8c5be356cd2873df9dfe61cc13f18bd00"} Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.875114 4926 generic.go:334] "Generic (PLEG): container finished" podID="4cb8f27f-336b-4acc-8165-0c53c1643084" containerID="6afc876661934d69ed6bb90885f853f8b5249b4339a13cab5b59071dc578e401" exitCode=0 Mar 12 18:24:20 crc kubenswrapper[4926]: I0312 18:24:20.875166 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-55g54" event={"ID":"4cb8f27f-336b-4acc-8165-0c53c1643084","Type":"ContainerDied","Data":"6afc876661934d69ed6bb90885f853f8b5249b4339a13cab5b59071dc578e401"} Mar 12 18:24:21 crc kubenswrapper[4926]: I0312 18:24:21.970882 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:24:21 crc kubenswrapper[4926]: I0312 18:24:21.971936 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fdd856968-nmxxn" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.089247 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6fc994c476-fv9c9"] Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.089778 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6fc994c476-fv9c9" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-log" containerID="cri-o://cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a" gracePeriod=30 Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.090713 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6fc994c476-fv9c9" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-api" containerID="cri-o://e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73" gracePeriod=30 Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.316283 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.438080 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxrpx\" (UniqueName: \"kubernetes.io/projected/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-kube-api-access-mxrpx\") pod \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.438304 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-operator-scripts\") pod \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\" (UID: \"dc4b3e47-ab9e-46c0-be37-9d2e94721be6\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.439838 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc4b3e47-ab9e-46c0-be37-9d2e94721be6" (UID: "dc4b3e47-ab9e-46c0-be37-9d2e94721be6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.444120 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-kube-api-access-mxrpx" (OuterVolumeSpecName: "kube-api-access-mxrpx") pod "dc4b3e47-ab9e-46c0-be37-9d2e94721be6" (UID: "dc4b3e47-ab9e-46c0-be37-9d2e94721be6"). InnerVolumeSpecName "kube-api-access-mxrpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.542585 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxrpx\" (UniqueName: \"kubernetes.io/projected/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-kube-api-access-mxrpx\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.542621 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc4b3e47-ab9e-46c0-be37-9d2e94721be6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.553176 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.566734 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.585250 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.604290 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.620059 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646538 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm9pf\" (UniqueName: \"kubernetes.io/projected/4cb8f27f-336b-4acc-8165-0c53c1643084-kube-api-access-jm9pf\") pod \"4cb8f27f-336b-4acc-8165-0c53c1643084\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646621 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rll4\" (UniqueName: \"kubernetes.io/projected/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-kube-api-access-5rll4\") pod \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646665 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbjlt\" (UniqueName: \"kubernetes.io/projected/15db9185-0441-431d-98db-5701f9b244be-kube-api-access-gbjlt\") pod \"15db9185-0441-431d-98db-5701f9b244be\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646727 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15db9185-0441-431d-98db-5701f9b244be-operator-scripts\") pod \"15db9185-0441-431d-98db-5701f9b244be\" (UID: \"15db9185-0441-431d-98db-5701f9b244be\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646757 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjfpj\" (UniqueName: \"kubernetes.io/projected/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-kube-api-access-hjfpj\") pod \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646787 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb8f27f-336b-4acc-8165-0c53c1643084-operator-scripts\") pod \"4cb8f27f-336b-4acc-8165-0c53c1643084\" (UID: \"4cb8f27f-336b-4acc-8165-0c53c1643084\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646835 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-operator-scripts\") pod \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\" (UID: \"fe0c0d35-e463-4b5c-a15f-cdfc808c498c\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.646867 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-operator-scripts\") pod \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\" (UID: \"b33ff126-1b0c-47f0-a4c1-18e1297fa81d\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.647670 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b33ff126-1b0c-47f0-a4c1-18e1297fa81d" (UID: "b33ff126-1b0c-47f0-a4c1-18e1297fa81d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.648425 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15db9185-0441-431d-98db-5701f9b244be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15db9185-0441-431d-98db-5701f9b244be" (UID: "15db9185-0441-431d-98db-5701f9b244be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.650929 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cb8f27f-336b-4acc-8165-0c53c1643084-kube-api-access-jm9pf" (OuterVolumeSpecName: "kube-api-access-jm9pf") pod "4cb8f27f-336b-4acc-8165-0c53c1643084" (UID: "4cb8f27f-336b-4acc-8165-0c53c1643084"). InnerVolumeSpecName "kube-api-access-jm9pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.651692 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-kube-api-access-5rll4" (OuterVolumeSpecName: "kube-api-access-5rll4") pod "b33ff126-1b0c-47f0-a4c1-18e1297fa81d" (UID: "b33ff126-1b0c-47f0-a4c1-18e1297fa81d"). InnerVolumeSpecName "kube-api-access-5rll4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.654587 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15db9185-0441-431d-98db-5701f9b244be-kube-api-access-gbjlt" (OuterVolumeSpecName: "kube-api-access-gbjlt") pod "15db9185-0441-431d-98db-5701f9b244be" (UID: "15db9185-0441-431d-98db-5701f9b244be"). InnerVolumeSpecName "kube-api-access-gbjlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.655363 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cb8f27f-336b-4acc-8165-0c53c1643084-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4cb8f27f-336b-4acc-8165-0c53c1643084" (UID: "4cb8f27f-336b-4acc-8165-0c53c1643084"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.655722 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe0c0d35-e463-4b5c-a15f-cdfc808c498c" (UID: "fe0c0d35-e463-4b5c-a15f-cdfc808c498c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.665620 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-kube-api-access-hjfpj" (OuterVolumeSpecName: "kube-api-access-hjfpj") pod "fe0c0d35-e463-4b5c-a15f-cdfc808c498c" (UID: "fe0c0d35-e463-4b5c-a15f-cdfc808c498c"). InnerVolumeSpecName "kube-api-access-hjfpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.748686 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29a2618b-1e0a-4246-b63c-582d2fe2f847-operator-scripts\") pod \"29a2618b-1e0a-4246-b63c-582d2fe2f847\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.748917 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8qwj\" (UniqueName: \"kubernetes.io/projected/29a2618b-1e0a-4246-b63c-582d2fe2f847-kube-api-access-p8qwj\") pod \"29a2618b-1e0a-4246-b63c-582d2fe2f847\" (UID: \"29a2618b-1e0a-4246-b63c-582d2fe2f847\") " Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749391 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm9pf\" (UniqueName: \"kubernetes.io/projected/4cb8f27f-336b-4acc-8165-0c53c1643084-kube-api-access-jm9pf\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749416 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rll4\" (UniqueName: \"kubernetes.io/projected/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-kube-api-access-5rll4\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749430 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbjlt\" (UniqueName: \"kubernetes.io/projected/15db9185-0441-431d-98db-5701f9b244be-kube-api-access-gbjlt\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749461 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15db9185-0441-431d-98db-5701f9b244be-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749472 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjfpj\" (UniqueName: \"kubernetes.io/projected/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-kube-api-access-hjfpj\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749485 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4cb8f27f-336b-4acc-8165-0c53c1643084-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749497 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe0c0d35-e463-4b5c-a15f-cdfc808c498c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.749510 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b33ff126-1b0c-47f0-a4c1-18e1297fa81d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.750798 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29a2618b-1e0a-4246-b63c-582d2fe2f847-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "29a2618b-1e0a-4246-b63c-582d2fe2f847" (UID: "29a2618b-1e0a-4246-b63c-582d2fe2f847"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.756635 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a2618b-1e0a-4246-b63c-582d2fe2f847-kube-api-access-p8qwj" (OuterVolumeSpecName: "kube-api-access-p8qwj") pod "29a2618b-1e0a-4246-b63c-582d2fe2f847" (UID: "29a2618b-1e0a-4246-b63c-582d2fe2f847"). InnerVolumeSpecName "kube-api-access-p8qwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.851112 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8qwj\" (UniqueName: \"kubernetes.io/projected/29a2618b-1e0a-4246-b63c-582d2fe2f847-kube-api-access-p8qwj\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.851341 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29a2618b-1e0a-4246-b63c-582d2fe2f847-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.907029 4926 generic.go:334] "Generic (PLEG): container finished" podID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerID="cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a" exitCode=143 Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.907106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fc994c476-fv9c9" event={"ID":"450e1ecf-5ae7-48b5-b567-e530e254f673","Type":"ContainerDied","Data":"cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.910141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerStarted","Data":"d0d68ef12f3c48e2e671f3c6b6e0ef4adbaa0a1069ddce15c7818c9465fd1365"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.910376 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.914559 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-382f-account-create-update-7sn9z" event={"ID":"fe0c0d35-e463-4b5c-a15f-cdfc808c498c","Type":"ContainerDied","Data":"b1d04c89d3d9fb6caf62bdb2b456380c91864ceab753f46a51bad242a32a7521"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.914604 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1d04c89d3d9fb6caf62bdb2b456380c91864ceab753f46a51bad242a32a7521" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.914682 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-382f-account-create-update-7sn9z" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.920857 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" event={"ID":"b33ff126-1b0c-47f0-a4c1-18e1297fa81d","Type":"ContainerDied","Data":"198fc93efd6dbc2346260a021b2acdc8c5be356cd2873df9dfe61cc13f18bd00"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.920895 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="198fc93efd6dbc2346260a021b2acdc8c5be356cd2873df9dfe61cc13f18bd00" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.920878 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f45b-account-create-update-rt9j9" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.922687 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-55g54" event={"ID":"4cb8f27f-336b-4acc-8165-0c53c1643084","Type":"ContainerDied","Data":"c6f0363211c7d4e74e3ccbba5ec4240749e8ebe5592c2510b2fa60c1f08ec3c9"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.922710 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-55g54" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.922724 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6f0363211c7d4e74e3ccbba5ec4240749e8ebe5592c2510b2fa60c1f08ec3c9" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.927044 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-0630-account-create-update-knx9n" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.927050 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-0630-account-create-update-knx9n" event={"ID":"29a2618b-1e0a-4246-b63c-582d2fe2f847","Type":"ContainerDied","Data":"ff34501c9af340574663fc5763d61f95a310be34a280df0b21b236386141ee47"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.927766 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff34501c9af340574663fc5763d61f95a310be34a280df0b21b236386141ee47" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.936822 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.501118291 podStartE2EDuration="6.936805548s" podCreationTimestamp="2026-03-12 18:24:16 +0000 UTC" firstStartedPulling="2026-03-12 18:24:17.588843105 +0000 UTC m=+1297.957469438" lastFinishedPulling="2026-03-12 18:24:22.024530362 +0000 UTC m=+1302.393156695" observedRunningTime="2026-03-12 18:24:22.932904356 +0000 UTC m=+1303.301530689" watchObservedRunningTime="2026-03-12 18:24:22.936805548 +0000 UTC m=+1303.305431871" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.938229 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-cctxg" event={"ID":"dc4b3e47-ab9e-46c0-be37-9d2e94721be6","Type":"ContainerDied","Data":"e0b512c7b0ec080f65f1af63e533a0195e053d720c63ac72bbcdd0918ebd9be0"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.938269 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0b512c7b0ec080f65f1af63e533a0195e053d720c63ac72bbcdd0918ebd9be0" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.938285 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-cctxg" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.943275 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-t4krn" Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.943596 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-t4krn" event={"ID":"15db9185-0441-431d-98db-5701f9b244be","Type":"ContainerDied","Data":"6c9449dc088a16c50ac498f087e95dd4074e327717458b2a2dca40a15417614a"} Mar 12 18:24:22 crc kubenswrapper[4926]: I0312 18:24:22.943720 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c9449dc088a16c50ac498f087e95dd4074e327717458b2a2dca40a15417614a" Mar 12 18:24:23 crc kubenswrapper[4926]: I0312 18:24:23.935958 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:24:23 crc kubenswrapper[4926]: I0312 18:24:23.940724 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-httpd" containerID="cri-o://1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178" gracePeriod=30 Mar 12 18:24:23 crc kubenswrapper[4926]: I0312 18:24:23.940682 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-log" containerID="cri-o://9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c" gracePeriod=30 Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.159157 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cbqv8"] Mar 12 18:24:24 crc kubenswrapper[4926]: E0312 18:24:24.159654 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cb8f27f-336b-4acc-8165-0c53c1643084" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.159680 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cb8f27f-336b-4acc-8165-0c53c1643084" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: E0312 18:24:24.159698 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15db9185-0441-431d-98db-5701f9b244be" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.159707 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="15db9185-0441-431d-98db-5701f9b244be" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: E0312 18:24:24.159720 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b33ff126-1b0c-47f0-a4c1-18e1297fa81d" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.159728 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b33ff126-1b0c-47f0-a4c1-18e1297fa81d" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: E0312 18:24:24.159739 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe0c0d35-e463-4b5c-a15f-cdfc808c498c" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.159747 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe0c0d35-e463-4b5c-a15f-cdfc808c498c" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: E0312 18:24:24.159774 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a2618b-1e0a-4246-b63c-582d2fe2f847" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.159782 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a2618b-1e0a-4246-b63c-582d2fe2f847" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: E0312 18:24:24.159810 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4b3e47-ab9e-46c0-be37-9d2e94721be6" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.159818 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4b3e47-ab9e-46c0-be37-9d2e94721be6" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.160030 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cb8f27f-336b-4acc-8165-0c53c1643084" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.160046 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b33ff126-1b0c-47f0-a4c1-18e1297fa81d" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.160060 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a2618b-1e0a-4246-b63c-582d2fe2f847" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.160077 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="15db9185-0441-431d-98db-5701f9b244be" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.160099 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe0c0d35-e463-4b5c-a15f-cdfc808c498c" containerName="mariadb-account-create-update" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.160114 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4b3e47-ab9e-46c0-be37-9d2e94721be6" containerName="mariadb-database-create" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.162506 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.171269 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.171710 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.171993 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-89pqc" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.186049 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cbqv8"] Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.275824 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-config-data\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.276497 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-scripts\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.276598 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbq59\" (UniqueName: \"kubernetes.io/projected/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-kube-api-access-fbq59\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.276688 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.378347 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbq59\" (UniqueName: \"kubernetes.io/projected/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-kube-api-access-fbq59\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.378615 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.378717 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-config-data\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.378909 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-scripts\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.384430 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-config-data\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.384992 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-scripts\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.390490 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.415109 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbq59\" (UniqueName: \"kubernetes.io/projected/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-kube-api-access-fbq59\") pod \"nova-cell0-conductor-db-sync-cbqv8\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.481060 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.804104 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cbqv8"] Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.983678 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" event={"ID":"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57","Type":"ContainerStarted","Data":"f726c742045749f4782471f8e9d8e4c6886b24fac2346f0142597eae7e927832"} Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.985554 4926 generic.go:334] "Generic (PLEG): container finished" podID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerID="9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c" exitCode=143 Mar 12 18:24:24 crc kubenswrapper[4926]: I0312 18:24:24.985590 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"27561f0e-1da4-4313-a7df-544fdfc893b1","Type":"ContainerDied","Data":"9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c"} Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.705464 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.725652 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-public-tls-certs\") pod \"450e1ecf-5ae7-48b5-b567-e530e254f673\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.726054 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-scripts\") pod \"450e1ecf-5ae7-48b5-b567-e530e254f673\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.726094 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vh2nd\" (UniqueName: \"kubernetes.io/projected/450e1ecf-5ae7-48b5-b567-e530e254f673-kube-api-access-vh2nd\") pod \"450e1ecf-5ae7-48b5-b567-e530e254f673\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.726192 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/450e1ecf-5ae7-48b5-b567-e530e254f673-logs\") pod \"450e1ecf-5ae7-48b5-b567-e530e254f673\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.726236 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-config-data\") pod \"450e1ecf-5ae7-48b5-b567-e530e254f673\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.726322 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-internal-tls-certs\") pod \"450e1ecf-5ae7-48b5-b567-e530e254f673\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.726343 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-combined-ca-bundle\") pod \"450e1ecf-5ae7-48b5-b567-e530e254f673\" (UID: \"450e1ecf-5ae7-48b5-b567-e530e254f673\") " Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.728124 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/450e1ecf-5ae7-48b5-b567-e530e254f673-logs" (OuterVolumeSpecName: "logs") pod "450e1ecf-5ae7-48b5-b567-e530e254f673" (UID: "450e1ecf-5ae7-48b5-b567-e530e254f673"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.737130 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-scripts" (OuterVolumeSpecName: "scripts") pod "450e1ecf-5ae7-48b5-b567-e530e254f673" (UID: "450e1ecf-5ae7-48b5-b567-e530e254f673"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.745638 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/450e1ecf-5ae7-48b5-b567-e530e254f673-kube-api-access-vh2nd" (OuterVolumeSpecName: "kube-api-access-vh2nd") pod "450e1ecf-5ae7-48b5-b567-e530e254f673" (UID: "450e1ecf-5ae7-48b5-b567-e530e254f673"). InnerVolumeSpecName "kube-api-access-vh2nd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.805957 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-config-data" (OuterVolumeSpecName: "config-data") pod "450e1ecf-5ae7-48b5-b567-e530e254f673" (UID: "450e1ecf-5ae7-48b5-b567-e530e254f673"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.829402 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/450e1ecf-5ae7-48b5-b567-e530e254f673-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.829449 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.829462 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.829475 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vh2nd\" (UniqueName: \"kubernetes.io/projected/450e1ecf-5ae7-48b5-b567-e530e254f673-kube-api-access-vh2nd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.839507 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "450e1ecf-5ae7-48b5-b567-e530e254f673" (UID: "450e1ecf-5ae7-48b5-b567-e530e254f673"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.856581 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "450e1ecf-5ae7-48b5-b567-e530e254f673" (UID: "450e1ecf-5ae7-48b5-b567-e530e254f673"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.870081 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "450e1ecf-5ae7-48b5-b567-e530e254f673" (UID: "450e1ecf-5ae7-48b5-b567-e530e254f673"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.930933 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.930998 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:25 crc kubenswrapper[4926]: I0312 18:24:25.931011 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/450e1ecf-5ae7-48b5-b567-e530e254f673-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.007999 4926 generic.go:334] "Generic (PLEG): container finished" podID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerID="e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73" exitCode=0 Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.008079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fc994c476-fv9c9" event={"ID":"450e1ecf-5ae7-48b5-b567-e530e254f673","Type":"ContainerDied","Data":"e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73"} Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.008115 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fc994c476-fv9c9" event={"ID":"450e1ecf-5ae7-48b5-b567-e530e254f673","Type":"ContainerDied","Data":"1f1ee7c53345b2a1183a515645bed6bfd508f8efad3f3b4a2588d4c7787c2e09"} Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.008387 4926 scope.go:117] "RemoveContainer" containerID="e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.008608 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fc994c476-fv9c9" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.051702 4926 scope.go:117] "RemoveContainer" containerID="cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.053464 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6fc994c476-fv9c9"] Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.063409 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6fc994c476-fv9c9"] Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.073818 4926 scope.go:117] "RemoveContainer" containerID="e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73" Mar 12 18:24:26 crc kubenswrapper[4926]: E0312 18:24:26.074189 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73\": container with ID starting with e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73 not found: ID does not exist" containerID="e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.074231 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73"} err="failed to get container status \"e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73\": rpc error: code = NotFound desc = could not find container \"e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73\": container with ID starting with e87064db3c4dab601e171b8ef221f01b14584cedfda748a9cbf2db9f357b6a73 not found: ID does not exist" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.074256 4926 scope.go:117] "RemoveContainer" containerID="cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a" Mar 12 18:24:26 crc kubenswrapper[4926]: E0312 18:24:26.077991 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a\": container with ID starting with cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a not found: ID does not exist" containerID="cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.078019 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a"} err="failed to get container status \"cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a\": rpc error: code = NotFound desc = could not find container \"cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a\": container with ID starting with cafbe824f3deda5f6c1263207977f209e8ce8e8b6bb8c642a6b34e03c9ace92a not found: ID does not exist" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.482960 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.488127 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-httpd" containerID="cri-o://73938b8cf4a571a4406931a492f2e033c549c10b7a2bcdf0ca6b902ede59649d" gracePeriod=30 Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.491503 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-log" containerID="cri-o://809f7347552b8f928619d96194c02f26e8cc98b9b07e8efde999fc4f99c7437f" gracePeriod=30 Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.513729 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" path="/var/lib/kubelet/pods/450e1ecf-5ae7-48b5-b567-e530e254f673/volumes" Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.560305 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.560570 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-central-agent" containerID="cri-o://70089f60f5708dc42a17c61910ff76a140c92fff334f5b497df2b30ef8db681d" gracePeriod=30 Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.560952 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="proxy-httpd" containerID="cri-o://d0d68ef12f3c48e2e671f3c6b6e0ef4adbaa0a1069ddce15c7818c9465fd1365" gracePeriod=30 Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.561003 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="sg-core" containerID="cri-o://ef9ed90f72e8dfcc270f36331c95e42d9aa6057e9d8ae052ebb5501d32f7648b" gracePeriod=30 Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.561035 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-notification-agent" containerID="cri-o://309dce8d3efcb58229ff7be17feec841c39fb231db3ee8c9eaa9375400b8f496" gracePeriod=30 Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.817349 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:24:26 crc kubenswrapper[4926]: I0312 18:24:26.817450 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.026137 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerID="d0d68ef12f3c48e2e671f3c6b6e0ef4adbaa0a1069ddce15c7818c9465fd1365" exitCode=0 Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.026173 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerID="ef9ed90f72e8dfcc270f36331c95e42d9aa6057e9d8ae052ebb5501d32f7648b" exitCode=2 Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.026181 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerID="309dce8d3efcb58229ff7be17feec841c39fb231db3ee8c9eaa9375400b8f496" exitCode=0 Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.026239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerDied","Data":"d0d68ef12f3c48e2e671f3c6b6e0ef4adbaa0a1069ddce15c7818c9465fd1365"} Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.026292 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerDied","Data":"ef9ed90f72e8dfcc270f36331c95e42d9aa6057e9d8ae052ebb5501d32f7648b"} Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.026314 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerDied","Data":"309dce8d3efcb58229ff7be17feec841c39fb231db3ee8c9eaa9375400b8f496"} Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.028874 4926 generic.go:334] "Generic (PLEG): container finished" podID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerID="809f7347552b8f928619d96194c02f26e8cc98b9b07e8efde999fc4f99c7437f" exitCode=143 Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.028922 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a","Type":"ContainerDied","Data":"809f7347552b8f928619d96194c02f26e8cc98b9b07e8efde999fc4f99c7437f"} Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.635066 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.760734 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.760796 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-scripts\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.760891 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9v2m5\" (UniqueName: \"kubernetes.io/projected/27561f0e-1da4-4313-a7df-544fdfc893b1-kube-api-access-9v2m5\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.760960 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-logs\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.761588 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-public-tls-certs\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.761622 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-httpd-run\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.761637 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-combined-ca-bundle\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.761719 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-config-data\") pod \"27561f0e-1da4-4313-a7df-544fdfc893b1\" (UID: \"27561f0e-1da4-4313-a7df-544fdfc893b1\") " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.761866 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.761934 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-logs" (OuterVolumeSpecName: "logs") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.762484 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.762499 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27561f0e-1da4-4313-a7df-544fdfc893b1-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.767020 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-scripts" (OuterVolumeSpecName: "scripts") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.768577 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.789895 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27561f0e-1da4-4313-a7df-544fdfc893b1-kube-api-access-9v2m5" (OuterVolumeSpecName: "kube-api-access-9v2m5") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "kube-api-access-9v2m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.805667 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.831879 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-config-data" (OuterVolumeSpecName: "config-data") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.842988 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "27561f0e-1da4-4313-a7df-544fdfc893b1" (UID: "27561f0e-1da4-4313-a7df-544fdfc893b1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.864209 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.864254 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.864273 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9v2m5\" (UniqueName: \"kubernetes.io/projected/27561f0e-1da4-4313-a7df-544fdfc893b1-kube-api-access-9v2m5\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.864289 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.864303 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.864315 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27561f0e-1da4-4313-a7df-544fdfc893b1-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.895522 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Mar 12 18:24:27 crc kubenswrapper[4926]: I0312 18:24:27.965564 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.041248 4926 generic.go:334] "Generic (PLEG): container finished" podID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerID="1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178" exitCode=0 Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.041319 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"27561f0e-1da4-4313-a7df-544fdfc893b1","Type":"ContainerDied","Data":"1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178"} Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.041361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"27561f0e-1da4-4313-a7df-544fdfc893b1","Type":"ContainerDied","Data":"600cdc83756bdbe2006c6d01ba6698b38db80c94206722c7e9c45bd9b1332b71"} Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.041386 4926 scope.go:117] "RemoveContainer" containerID="1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.041594 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.083745 4926 scope.go:117] "RemoveContainer" containerID="9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.092160 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.111920 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.126339 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:24:28 crc kubenswrapper[4926]: E0312 18:24:28.129231 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-api" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.129289 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-api" Mar 12 18:24:28 crc kubenswrapper[4926]: E0312 18:24:28.129524 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-log" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.129538 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-log" Mar 12 18:24:28 crc kubenswrapper[4926]: E0312 18:24:28.129567 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-httpd" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.129761 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-httpd" Mar 12 18:24:28 crc kubenswrapper[4926]: E0312 18:24:28.129787 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-log" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.129796 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-log" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.132020 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-log" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.132050 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="450e1ecf-5ae7-48b5-b567-e530e254f673" containerName="placement-api" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.132072 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-log" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.132102 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" containerName="glance-httpd" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.133205 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.138173 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.139652 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.141229 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.177725 4926 scope.go:117] "RemoveContainer" containerID="1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178" Mar 12 18:24:28 crc kubenswrapper[4926]: E0312 18:24:28.179868 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178\": container with ID starting with 1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178 not found: ID does not exist" containerID="1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.179918 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178"} err="failed to get container status \"1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178\": rpc error: code = NotFound desc = could not find container \"1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178\": container with ID starting with 1fa3d22484b03763d8133f2d587a25b5e1ffc1e16aa1a1f610d78f1e559cc178 not found: ID does not exist" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.179964 4926 scope.go:117] "RemoveContainer" containerID="9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c" Mar 12 18:24:28 crc kubenswrapper[4926]: E0312 18:24:28.181846 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c\": container with ID starting with 9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c not found: ID does not exist" containerID="9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.181914 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c"} err="failed to get container status \"9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c\": rpc error: code = NotFound desc = could not find container \"9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c\": container with ID starting with 9c7305029bac56d73a118acb64bb658e6027d0c4477cedae842b84b08298a73c not found: ID does not exist" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5lxj\" (UniqueName: \"kubernetes.io/projected/0749ce08-7fa5-48fe-9248-ac3a6699ef57-kube-api-access-z5lxj\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269751 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-scripts\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269788 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0749ce08-7fa5-48fe-9248-ac3a6699ef57-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269852 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-config-data\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269894 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269916 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269938 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.269970 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0749ce08-7fa5-48fe-9248-ac3a6699ef57-logs\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.371893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5lxj\" (UniqueName: \"kubernetes.io/projected/0749ce08-7fa5-48fe-9248-ac3a6699ef57-kube-api-access-z5lxj\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.371938 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-scripts\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.371964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0749ce08-7fa5-48fe-9248-ac3a6699ef57-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372016 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-config-data\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372053 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372089 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372116 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0749ce08-7fa5-48fe-9248-ac3a6699ef57-logs\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372686 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0749ce08-7fa5-48fe-9248-ac3a6699ef57-logs\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372828 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0749ce08-7fa5-48fe-9248-ac3a6699ef57-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.372927 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.377972 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.378370 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.378844 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-scripts\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.379350 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0749ce08-7fa5-48fe-9248-ac3a6699ef57-config-data\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.399422 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5lxj\" (UniqueName: \"kubernetes.io/projected/0749ce08-7fa5-48fe-9248-ac3a6699ef57-kube-api-access-z5lxj\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.421515 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0749ce08-7fa5-48fe-9248-ac3a6699ef57\") " pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.500422 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 12 18:24:28 crc kubenswrapper[4926]: I0312 18:24:28.501286 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27561f0e-1da4-4313-a7df-544fdfc893b1" path="/var/lib/kubelet/pods/27561f0e-1da4-4313-a7df-544fdfc893b1/volumes" Mar 12 18:24:29 crc kubenswrapper[4926]: I0312 18:24:29.087982 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 12 18:24:29 crc kubenswrapper[4926]: I0312 18:24:29.643526 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:53948->10.217.0.154:9292: read: connection reset by peer" Mar 12 18:24:29 crc kubenswrapper[4926]: I0312 18:24:29.644405 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.154:9292/healthcheck\": read tcp 10.217.0.2:53946->10.217.0.154:9292: read: connection reset by peer" Mar 12 18:24:30 crc kubenswrapper[4926]: I0312 18:24:30.069215 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerID="70089f60f5708dc42a17c61910ff76a140c92fff334f5b497df2b30ef8db681d" exitCode=0 Mar 12 18:24:30 crc kubenswrapper[4926]: I0312 18:24:30.069271 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerDied","Data":"70089f60f5708dc42a17c61910ff76a140c92fff334f5b497df2b30ef8db681d"} Mar 12 18:24:30 crc kubenswrapper[4926]: I0312 18:24:30.071045 4926 generic.go:334] "Generic (PLEG): container finished" podID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerID="73938b8cf4a571a4406931a492f2e033c549c10b7a2bcdf0ca6b902ede59649d" exitCode=0 Mar 12 18:24:30 crc kubenswrapper[4926]: I0312 18:24:30.071072 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a","Type":"ContainerDied","Data":"73938b8cf4a571a4406931a492f2e033c549c10b7a2bcdf0ca6b902ede59649d"} Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.108824 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0749ce08-7fa5-48fe-9248-ac3a6699ef57","Type":"ContainerStarted","Data":"a4a12cabad0ea532415d393ea0ffd34bbf6b2b5705b28bea1f5b26dd3d7804de"} Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.442028 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.486632 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505106 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-httpd-run\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505164 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-config-data\") pod \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505213 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-scripts\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505237 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-scripts\") pod \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505260 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-config-data\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505353 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-run-httpd\") pod \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505382 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-log-httpd\") pod \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505407 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qgwd\" (UniqueName: \"kubernetes.io/projected/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-kube-api-access-7qgwd\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505430 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-internal-tls-certs\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505887 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-combined-ca-bundle\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.505983 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-combined-ca-bundle\") pod \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506164 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" (UID: "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506336 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b9v4\" (UniqueName: \"kubernetes.io/projected/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-kube-api-access-7b9v4\") pod \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506387 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-logs\") pod \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\" (UID: \"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506416 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-sg-core-conf-yaml\") pod \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\" (UID: \"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea\") " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506479 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" (UID: "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506910 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-logs" (OuterVolumeSpecName: "logs") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506976 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.506995 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.507004 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.509860 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-kube-api-access-7qgwd" (OuterVolumeSpecName: "kube-api-access-7qgwd") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "kube-api-access-7qgwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.511261 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.539251 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-scripts" (OuterVolumeSpecName: "scripts") pod "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" (UID: "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.539271 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-scripts" (OuterVolumeSpecName: "scripts") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.540329 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-kube-api-access-7b9v4" (OuterVolumeSpecName: "kube-api-access-7b9v4") pod "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" (UID: "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea"). InnerVolumeSpecName "kube-api-access-7b9v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.567264 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" (UID: "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.583744 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.594241 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608614 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b9v4\" (UniqueName: \"kubernetes.io/projected/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-kube-api-access-7b9v4\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608644 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608656 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608663 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608671 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608679 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qgwd\" (UniqueName: \"kubernetes.io/projected/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-kube-api-access-7qgwd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608687 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608712 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.608721 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.614094 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-config-data" (OuterVolumeSpecName: "config-data") pod "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" (UID: "f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.628062 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.633404 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-config-data" (OuterVolumeSpecName: "config-data") pod "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" (UID: "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.673192 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" (UID: "2e50e6cb-b07b-4a13-8b5c-e889e53c49ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.709804 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.710048 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.710060 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:34 crc kubenswrapper[4926]: I0312 18:24:34.710070 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.125178 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e50e6cb-b07b-4a13-8b5c-e889e53c49ea","Type":"ContainerDied","Data":"7996e6fe65ac453b2bf7a0036550739c5607aa7bde6261bcc4407d93c97a4ca7"} Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.125233 4926 scope.go:117] "RemoveContainer" containerID="d0d68ef12f3c48e2e671f3c6b6e0ef4adbaa0a1069ddce15c7818c9465fd1365" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.125285 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.132363 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.132480 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a","Type":"ContainerDied","Data":"ac524b6af20000079258195bab5b0ef6d7a0ae5f18a8830578a4f4e1b53a9c85"} Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.136239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0749ce08-7fa5-48fe-9248-ac3a6699ef57","Type":"ContainerStarted","Data":"157de7d9a24e5910cf8461e4d788dc27f9a81db1355cf4f0c8d758eeeec9fd24"} Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.138364 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" event={"ID":"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57","Type":"ContainerStarted","Data":"683758cbf77fef46bca458acf7d282b0867f1aa1a065fbc9925548f2b1c73316"} Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.173099 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" podStartSLOduration=1.920938977 podStartE2EDuration="11.173077489s" podCreationTimestamp="2026-03-12 18:24:24 +0000 UTC" firstStartedPulling="2026-03-12 18:24:24.820245218 +0000 UTC m=+1305.188871551" lastFinishedPulling="2026-03-12 18:24:34.07238373 +0000 UTC m=+1314.441010063" observedRunningTime="2026-03-12 18:24:35.157710336 +0000 UTC m=+1315.526336689" watchObservedRunningTime="2026-03-12 18:24:35.173077489 +0000 UTC m=+1315.541703822" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.180579 4926 scope.go:117] "RemoveContainer" containerID="ef9ed90f72e8dfcc270f36331c95e42d9aa6057e9d8ae052ebb5501d32f7648b" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.203399 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.212735 4926 scope.go:117] "RemoveContainer" containerID="309dce8d3efcb58229ff7be17feec841c39fb231db3ee8c9eaa9375400b8f496" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.221890 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.249938 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.255969 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: E0312 18:24:35.256806 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-log" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.256832 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-log" Mar 12 18:24:35 crc kubenswrapper[4926]: E0312 18:24:35.256849 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="sg-core" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.256858 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="sg-core" Mar 12 18:24:35 crc kubenswrapper[4926]: E0312 18:24:35.256882 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-httpd" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.256892 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-httpd" Mar 12 18:24:35 crc kubenswrapper[4926]: E0312 18:24:35.256908 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-central-agent" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.256915 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-central-agent" Mar 12 18:24:35 crc kubenswrapper[4926]: E0312 18:24:35.256934 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-notification-agent" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.256942 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-notification-agent" Mar 12 18:24:35 crc kubenswrapper[4926]: E0312 18:24:35.256956 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="proxy-httpd" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.256963 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="proxy-httpd" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.257158 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-log" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.257183 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-central-agent" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.257202 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="proxy-httpd" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.257216 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="sg-core" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.257229 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" containerName="ceilometer-notification-agent" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.257240 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" containerName="glance-httpd" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.258481 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.263170 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.263243 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.263318 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.264550 4926 scope.go:117] "RemoveContainer" containerID="70089f60f5708dc42a17c61910ff76a140c92fff334f5b497df2b30ef8db681d" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.284556 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.292266 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.294246 4926 scope.go:117] "RemoveContainer" containerID="73938b8cf4a571a4406931a492f2e033c549c10b7a2bcdf0ca6b902ede59649d" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.294702 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.298773 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.313300 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.337531 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.340035 4926 scope.go:117] "RemoveContainer" containerID="809f7347552b8f928619d96194c02f26e8cc98b9b07e8efde999fc4f99c7437f" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428311 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428365 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-run-httpd\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428416 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428493 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqxnc\" (UniqueName: \"kubernetes.io/projected/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-kube-api-access-bqxnc\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428529 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428554 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428582 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-log-httpd\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428620 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428640 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-config-data\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428696 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428731 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.428852 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.429003 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq7gn\" (UniqueName: \"kubernetes.io/projected/f48005ff-c364-47aa-8253-9a3124097a10-kube-api-access-lq7gn\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.429045 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-scripts\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530558 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530610 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530652 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq7gn\" (UniqueName: \"kubernetes.io/projected/f48005ff-c364-47aa-8253-9a3124097a10-kube-api-access-lq7gn\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530673 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-scripts\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530702 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530717 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-run-httpd\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530748 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530783 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqxnc\" (UniqueName: \"kubernetes.io/projected/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-kube-api-access-bqxnc\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530804 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530844 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-log-httpd\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530869 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530886 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-config-data\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530900 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.530924 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.531370 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.532633 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-log-httpd\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.532664 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-run-httpd\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.532750 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.532944 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-logs\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.537109 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.537148 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.537301 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.537954 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-scripts\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.546178 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.548566 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.550043 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.550385 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-config-data\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.553331 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq7gn\" (UniqueName: \"kubernetes.io/projected/f48005ff-c364-47aa-8253-9a3124097a10-kube-api-access-lq7gn\") pod \"ceilometer-0\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.554127 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqxnc\" (UniqueName: \"kubernetes.io/projected/1e75a0d1-4272-43f3-b1b8-3cfe57e0141d-kube-api-access-bqxnc\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.579466 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d\") " pod="openstack/glance-default-internal-api-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.642223 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:35 crc kubenswrapper[4926]: I0312 18:24:35.877277 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:36 crc kubenswrapper[4926]: I0312 18:24:36.072428 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:36 crc kubenswrapper[4926]: W0312 18:24:36.075641 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf48005ff_c364_47aa_8253_9a3124097a10.slice/crio-f451adeb6303d8745b2e50f6138760aa3de2cccdedbd3b8de83c0f3d99a02905 WatchSource:0}: Error finding container f451adeb6303d8745b2e50f6138760aa3de2cccdedbd3b8de83c0f3d99a02905: Status 404 returned error can't find the container with id f451adeb6303d8745b2e50f6138760aa3de2cccdedbd3b8de83c0f3d99a02905 Mar 12 18:24:36 crc kubenswrapper[4926]: I0312 18:24:36.149494 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0749ce08-7fa5-48fe-9248-ac3a6699ef57","Type":"ContainerStarted","Data":"0422041cdf5f4f73ab75457856395019cfba45f8d3b4cb8fec90717ce08223b5"} Mar 12 18:24:36 crc kubenswrapper[4926]: I0312 18:24:36.152407 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerStarted","Data":"f451adeb6303d8745b2e50f6138760aa3de2cccdedbd3b8de83c0f3d99a02905"} Mar 12 18:24:36 crc kubenswrapper[4926]: I0312 18:24:36.181361 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.181342816 podStartE2EDuration="8.181342816s" podCreationTimestamp="2026-03-12 18:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:24:36.169146884 +0000 UTC m=+1316.537773247" watchObservedRunningTime="2026-03-12 18:24:36.181342816 +0000 UTC m=+1316.549969149" Mar 12 18:24:36 crc kubenswrapper[4926]: I0312 18:24:36.366532 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 12 18:24:36 crc kubenswrapper[4926]: W0312 18:24:36.367695 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e75a0d1_4272_43f3_b1b8_3cfe57e0141d.slice/crio-5ccda701c4d814f87501797d7fb8c8b92ceeefdf90d8b6cef3192f73df6c015a WatchSource:0}: Error finding container 5ccda701c4d814f87501797d7fb8c8b92ceeefdf90d8b6cef3192f73df6c015a: Status 404 returned error can't find the container with id 5ccda701c4d814f87501797d7fb8c8b92ceeefdf90d8b6cef3192f73df6c015a Mar 12 18:24:36 crc kubenswrapper[4926]: I0312 18:24:36.514349 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e50e6cb-b07b-4a13-8b5c-e889e53c49ea" path="/var/lib/kubelet/pods/2e50e6cb-b07b-4a13-8b5c-e889e53c49ea/volumes" Mar 12 18:24:36 crc kubenswrapper[4926]: I0312 18:24:36.515715 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a" path="/var/lib/kubelet/pods/f3945fa8-a1cc-4c4b-91c6-2cb5a6567f8a/volumes" Mar 12 18:24:37 crc kubenswrapper[4926]: I0312 18:24:37.167668 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerStarted","Data":"435ef2132df5554295889ec9ba808434afc58d13077a9a39e7222dd605149d55"} Mar 12 18:24:37 crc kubenswrapper[4926]: I0312 18:24:37.171164 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d","Type":"ContainerStarted","Data":"e57f630ce30cab4c95cd232834974e08eb8649298261ded7bcb7b40c4db45ae8"} Mar 12 18:24:37 crc kubenswrapper[4926]: I0312 18:24:37.171196 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d","Type":"ContainerStarted","Data":"5ccda701c4d814f87501797d7fb8c8b92ceeefdf90d8b6cef3192f73df6c015a"} Mar 12 18:24:38 crc kubenswrapper[4926]: I0312 18:24:38.183118 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerStarted","Data":"0b78f1753e8c4a22ed2f4147f06789b56f6db668b7a92626590701ea109bcd09"} Mar 12 18:24:38 crc kubenswrapper[4926]: I0312 18:24:38.185934 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1e75a0d1-4272-43f3-b1b8-3cfe57e0141d","Type":"ContainerStarted","Data":"42e78e74e76a93045e02e79927bae762527f964a9237f343e0e05360197b5f00"} Mar 12 18:24:38 crc kubenswrapper[4926]: I0312 18:24:38.212120 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.212099439 podStartE2EDuration="3.212099439s" podCreationTimestamp="2026-03-12 18:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:24:38.206879486 +0000 UTC m=+1318.575505819" watchObservedRunningTime="2026-03-12 18:24:38.212099439 +0000 UTC m=+1318.580725782" Mar 12 18:24:38 crc kubenswrapper[4926]: I0312 18:24:38.503728 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 18:24:38 crc kubenswrapper[4926]: I0312 18:24:38.504086 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 12 18:24:38 crc kubenswrapper[4926]: I0312 18:24:38.549702 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 18:24:38 crc kubenswrapper[4926]: I0312 18:24:38.553690 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 12 18:24:39 crc kubenswrapper[4926]: I0312 18:24:39.197891 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerStarted","Data":"7655ace271d6f3feff7f2f4eaa16c5ea625e5fe6beab0c4038907b7f5e715987"} Mar 12 18:24:39 crc kubenswrapper[4926]: I0312 18:24:39.198885 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 18:24:39 crc kubenswrapper[4926]: I0312 18:24:39.198964 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 12 18:24:39 crc kubenswrapper[4926]: I0312 18:24:39.425808 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.222479 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.222565 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-central-agent" containerID="cri-o://435ef2132df5554295889ec9ba808434afc58d13077a9a39e7222dd605149d55" gracePeriod=30 Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.222601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerStarted","Data":"89db61a8e375fc043c0de0eeea971f98c1fb8c205fa68c44c6f73887667808c6"} Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.222850 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="sg-core" containerID="cri-o://7655ace271d6f3feff7f2f4eaa16c5ea625e5fe6beab0c4038907b7f5e715987" gracePeriod=30 Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.223429 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.222892 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-notification-agent" containerID="cri-o://0b78f1753e8c4a22ed2f4147f06789b56f6db668b7a92626590701ea109bcd09" gracePeriod=30 Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.222879 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="proxy-httpd" containerID="cri-o://89db61a8e375fc043c0de0eeea971f98c1fb8c205fa68c44c6f73887667808c6" gracePeriod=30 Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.253216 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9285094919999999 podStartE2EDuration="6.253198906s" podCreationTimestamp="2026-03-12 18:24:35 +0000 UTC" firstStartedPulling="2026-03-12 18:24:36.078095456 +0000 UTC m=+1316.446721789" lastFinishedPulling="2026-03-12 18:24:40.40278484 +0000 UTC m=+1320.771411203" observedRunningTime="2026-03-12 18:24:41.244678318 +0000 UTC m=+1321.613304661" watchObservedRunningTime="2026-03-12 18:24:41.253198906 +0000 UTC m=+1321.621825239" Mar 12 18:24:41 crc kubenswrapper[4926]: I0312 18:24:41.307848 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 18:24:42 crc kubenswrapper[4926]: I0312 18:24:42.094060 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 12 18:24:42 crc kubenswrapper[4926]: I0312 18:24:42.236650 4926 generic.go:334] "Generic (PLEG): container finished" podID="f48005ff-c364-47aa-8253-9a3124097a10" containerID="89db61a8e375fc043c0de0eeea971f98c1fb8c205fa68c44c6f73887667808c6" exitCode=0 Mar 12 18:24:42 crc kubenswrapper[4926]: I0312 18:24:42.236684 4926 generic.go:334] "Generic (PLEG): container finished" podID="f48005ff-c364-47aa-8253-9a3124097a10" containerID="7655ace271d6f3feff7f2f4eaa16c5ea625e5fe6beab0c4038907b7f5e715987" exitCode=2 Mar 12 18:24:42 crc kubenswrapper[4926]: I0312 18:24:42.236692 4926 generic.go:334] "Generic (PLEG): container finished" podID="f48005ff-c364-47aa-8253-9a3124097a10" containerID="0b78f1753e8c4a22ed2f4147f06789b56f6db668b7a92626590701ea109bcd09" exitCode=0 Mar 12 18:24:42 crc kubenswrapper[4926]: I0312 18:24:42.237322 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerDied","Data":"89db61a8e375fc043c0de0eeea971f98c1fb8c205fa68c44c6f73887667808c6"} Mar 12 18:24:42 crc kubenswrapper[4926]: I0312 18:24:42.237378 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerDied","Data":"7655ace271d6f3feff7f2f4eaa16c5ea625e5fe6beab0c4038907b7f5e715987"} Mar 12 18:24:42 crc kubenswrapper[4926]: I0312 18:24:42.237393 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerDied","Data":"0b78f1753e8c4a22ed2f4147f06789b56f6db668b7a92626590701ea109bcd09"} Mar 12 18:24:45 crc kubenswrapper[4926]: I0312 18:24:45.878844 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:45 crc kubenswrapper[4926]: I0312 18:24:45.879632 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:45 crc kubenswrapper[4926]: I0312 18:24:45.920305 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:45 crc kubenswrapper[4926]: I0312 18:24:45.957123 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.276649 4926 generic.go:334] "Generic (PLEG): container finished" podID="436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" containerID="683758cbf77fef46bca458acf7d282b0867f1aa1a065fbc9925548f2b1c73316" exitCode=0 Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.276744 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" event={"ID":"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57","Type":"ContainerDied","Data":"683758cbf77fef46bca458acf7d282b0867f1aa1a065fbc9925548f2b1c73316"} Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.282113 4926 generic.go:334] "Generic (PLEG): container finished" podID="f48005ff-c364-47aa-8253-9a3124097a10" containerID="435ef2132df5554295889ec9ba808434afc58d13077a9a39e7222dd605149d55" exitCode=0 Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.282179 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerDied","Data":"435ef2132df5554295889ec9ba808434afc58d13077a9a39e7222dd605149d55"} Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.282225 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f48005ff-c364-47aa-8253-9a3124097a10","Type":"ContainerDied","Data":"f451adeb6303d8745b2e50f6138760aa3de2cccdedbd3b8de83c0f3d99a02905"} Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.282241 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f451adeb6303d8745b2e50f6138760aa3de2cccdedbd3b8de83c0f3d99a02905" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.282485 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.282520 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.297120 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.482176 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq7gn\" (UniqueName: \"kubernetes.io/projected/f48005ff-c364-47aa-8253-9a3124097a10-kube-api-access-lq7gn\") pod \"f48005ff-c364-47aa-8253-9a3124097a10\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.482247 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-config-data\") pod \"f48005ff-c364-47aa-8253-9a3124097a10\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.482425 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-combined-ca-bundle\") pod \"f48005ff-c364-47aa-8253-9a3124097a10\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.482513 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-scripts\") pod \"f48005ff-c364-47aa-8253-9a3124097a10\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.482548 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-sg-core-conf-yaml\") pod \"f48005ff-c364-47aa-8253-9a3124097a10\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.482636 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-run-httpd\") pod \"f48005ff-c364-47aa-8253-9a3124097a10\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.482669 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-log-httpd\") pod \"f48005ff-c364-47aa-8253-9a3124097a10\" (UID: \"f48005ff-c364-47aa-8253-9a3124097a10\") " Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.483326 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f48005ff-c364-47aa-8253-9a3124097a10" (UID: "f48005ff-c364-47aa-8253-9a3124097a10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.483581 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f48005ff-c364-47aa-8253-9a3124097a10" (UID: "f48005ff-c364-47aa-8253-9a3124097a10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.484046 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.484076 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f48005ff-c364-47aa-8253-9a3124097a10-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.488102 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f48005ff-c364-47aa-8253-9a3124097a10-kube-api-access-lq7gn" (OuterVolumeSpecName: "kube-api-access-lq7gn") pod "f48005ff-c364-47aa-8253-9a3124097a10" (UID: "f48005ff-c364-47aa-8253-9a3124097a10"). InnerVolumeSpecName "kube-api-access-lq7gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.488952 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-scripts" (OuterVolumeSpecName: "scripts") pod "f48005ff-c364-47aa-8253-9a3124097a10" (UID: "f48005ff-c364-47aa-8253-9a3124097a10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.514243 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f48005ff-c364-47aa-8253-9a3124097a10" (UID: "f48005ff-c364-47aa-8253-9a3124097a10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.561540 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f48005ff-c364-47aa-8253-9a3124097a10" (UID: "f48005ff-c364-47aa-8253-9a3124097a10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.586301 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.586351 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.586370 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.586390 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq7gn\" (UniqueName: \"kubernetes.io/projected/f48005ff-c364-47aa-8253-9a3124097a10-kube-api-access-lq7gn\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.602049 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-config-data" (OuterVolumeSpecName: "config-data") pod "f48005ff-c364-47aa-8253-9a3124097a10" (UID: "f48005ff-c364-47aa-8253-9a3124097a10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:46 crc kubenswrapper[4926]: I0312 18:24:46.688531 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f48005ff-c364-47aa-8253-9a3124097a10-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.291594 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.349164 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.358723 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.374527 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:47 crc kubenswrapper[4926]: E0312 18:24:47.374996 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="proxy-httpd" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375015 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="proxy-httpd" Mar 12 18:24:47 crc kubenswrapper[4926]: E0312 18:24:47.375039 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="sg-core" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375048 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="sg-core" Mar 12 18:24:47 crc kubenswrapper[4926]: E0312 18:24:47.375062 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-central-agent" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375070 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-central-agent" Mar 12 18:24:47 crc kubenswrapper[4926]: E0312 18:24:47.375084 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-notification-agent" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375091 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-notification-agent" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375287 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="proxy-httpd" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375306 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-central-agent" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375317 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="ceilometer-notification-agent" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.375330 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48005ff-c364-47aa-8253-9a3124097a10" containerName="sg-core" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.382057 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.384217 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.386801 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.392064 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.516480 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-scripts\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.516539 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-log-httpd\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.516566 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-config-data\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.516640 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-run-httpd\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.516662 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.516686 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.516710 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbct5\" (UniqueName: \"kubernetes.io/projected/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-kube-api-access-dbct5\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619028 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-log-httpd\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619103 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-config-data\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-run-httpd\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619429 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619549 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbct5\" (UniqueName: \"kubernetes.io/projected/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-kube-api-access-dbct5\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619614 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-scripts\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.619882 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-run-httpd\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.620197 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-log-httpd\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.625297 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.627151 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.628129 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-scripts\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.634084 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-config-data\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.641968 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbct5\" (UniqueName: \"kubernetes.io/projected/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-kube-api-access-dbct5\") pod \"ceilometer-0\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.707733 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.741884 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.935178 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbq59\" (UniqueName: \"kubernetes.io/projected/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-kube-api-access-fbq59\") pod \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.935207 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-scripts\") pod \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.935246 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-config-data\") pod \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.935310 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-combined-ca-bundle\") pod \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\" (UID: \"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57\") " Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.940566 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-kube-api-access-fbq59" (OuterVolumeSpecName: "kube-api-access-fbq59") pod "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" (UID: "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57"). InnerVolumeSpecName "kube-api-access-fbq59". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.944793 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-scripts" (OuterVolumeSpecName: "scripts") pod "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" (UID: "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.975989 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" (UID: "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:47 crc kubenswrapper[4926]: I0312 18:24:47.994386 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-config-data" (OuterVolumeSpecName: "config-data") pod "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" (UID: "436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.038653 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.038711 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbq59\" (UniqueName: \"kubernetes.io/projected/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-kube-api-access-fbq59\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.038733 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.038747 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.195530 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.303476 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerStarted","Data":"ab0418481e92c6856771c7ea5c23c736e2ec58cc43ca849b59c1411888deb2c2"} Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.305650 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.306550 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.312682 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cbqv8" event={"ID":"436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57","Type":"ContainerDied","Data":"f726c742045749f4782471f8e9d8e4c6886b24fac2346f0142597eae7e927832"} Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.312752 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f726c742045749f4782471f8e9d8e4c6886b24fac2346f0142597eae7e927832" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.312778 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.448068 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 18:24:48 crc kubenswrapper[4926]: E0312 18:24:48.448374 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" containerName="nova-cell0-conductor-db-sync" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.448386 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" containerName="nova-cell0-conductor-db-sync" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.448591 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" containerName="nova-cell0-conductor-db-sync" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.449126 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.452875 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.453060 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-89pqc" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.462683 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.500096 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f48005ff-c364-47aa-8253-9a3124097a10" path="/var/lib/kubelet/pods/f48005ff-c364-47aa-8253-9a3124097a10/volumes" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.551870 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397b3426-1e6c-468a-8d6d-562e70944d9d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.551918 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7d6b\" (UniqueName: \"kubernetes.io/projected/397b3426-1e6c-468a-8d6d-562e70944d9d-kube-api-access-h7d6b\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.552005 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397b3426-1e6c-468a-8d6d-562e70944d9d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.552747 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.655752 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397b3426-1e6c-468a-8d6d-562e70944d9d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.655820 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7d6b\" (UniqueName: \"kubernetes.io/projected/397b3426-1e6c-468a-8d6d-562e70944d9d-kube-api-access-h7d6b\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.655917 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397b3426-1e6c-468a-8d6d-562e70944d9d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.661287 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/397b3426-1e6c-468a-8d6d-562e70944d9d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.661732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/397b3426-1e6c-468a-8d6d-562e70944d9d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.686232 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7d6b\" (UniqueName: \"kubernetes.io/projected/397b3426-1e6c-468a-8d6d-562e70944d9d-kube-api-access-h7d6b\") pod \"nova-cell0-conductor-0\" (UID: \"397b3426-1e6c-468a-8d6d-562e70944d9d\") " pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:48 crc kubenswrapper[4926]: I0312 18:24:48.765877 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:49 crc kubenswrapper[4926]: I0312 18:24:49.233267 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 12 18:24:49 crc kubenswrapper[4926]: I0312 18:24:49.317602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"397b3426-1e6c-468a-8d6d-562e70944d9d","Type":"ContainerStarted","Data":"58b6f95ba06f065609956e1b292ef3638775b20aa0e44627ce4cf95bbe74986b"} Mar 12 18:24:49 crc kubenswrapper[4926]: I0312 18:24:49.327072 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerStarted","Data":"5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d"} Mar 12 18:24:50 crc kubenswrapper[4926]: I0312 18:24:50.336858 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerStarted","Data":"5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d"} Mar 12 18:24:50 crc kubenswrapper[4926]: I0312 18:24:50.339678 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"397b3426-1e6c-468a-8d6d-562e70944d9d","Type":"ContainerStarted","Data":"6389117384c859ad32657d06cff7727c2a26c4645a8bf2642376f03ed6f98468"} Mar 12 18:24:50 crc kubenswrapper[4926]: I0312 18:24:50.341066 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:50 crc kubenswrapper[4926]: I0312 18:24:50.361954 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.361934057 podStartE2EDuration="2.361934057s" podCreationTimestamp="2026-03-12 18:24:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:24:50.355522326 +0000 UTC m=+1330.724148669" watchObservedRunningTime="2026-03-12 18:24:50.361934057 +0000 UTC m=+1330.730560390" Mar 12 18:24:51 crc kubenswrapper[4926]: I0312 18:24:51.354687 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerStarted","Data":"26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b"} Mar 12 18:24:53 crc kubenswrapper[4926]: I0312 18:24:53.381192 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerStarted","Data":"271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0"} Mar 12 18:24:53 crc kubenswrapper[4926]: I0312 18:24:53.383496 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:24:53 crc kubenswrapper[4926]: I0312 18:24:53.418084 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.087379613 podStartE2EDuration="6.418066725s" podCreationTimestamp="2026-03-12 18:24:47 +0000 UTC" firstStartedPulling="2026-03-12 18:24:48.287153933 +0000 UTC m=+1328.655780276" lastFinishedPulling="2026-03-12 18:24:52.617841045 +0000 UTC m=+1332.986467388" observedRunningTime="2026-03-12 18:24:53.412141739 +0000 UTC m=+1333.780768072" watchObservedRunningTime="2026-03-12 18:24:53.418066725 +0000 UTC m=+1333.786693058" Mar 12 18:24:56 crc kubenswrapper[4926]: I0312 18:24:56.818121 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:24:56 crc kubenswrapper[4926]: I0312 18:24:56.818490 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:24:56 crc kubenswrapper[4926]: I0312 18:24:56.818559 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:24:56 crc kubenswrapper[4926]: I0312 18:24:56.819491 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9728bd6132bdd9ab31a71d0a44779a02f515c39e712bb7cc4f8a85610efe739f"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:24:56 crc kubenswrapper[4926]: I0312 18:24:56.819566 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://9728bd6132bdd9ab31a71d0a44779a02f515c39e712bb7cc4f8a85610efe739f" gracePeriod=600 Mar 12 18:24:57 crc kubenswrapper[4926]: I0312 18:24:57.435593 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="9728bd6132bdd9ab31a71d0a44779a02f515c39e712bb7cc4f8a85610efe739f" exitCode=0 Mar 12 18:24:57 crc kubenswrapper[4926]: I0312 18:24:57.435667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"9728bd6132bdd9ab31a71d0a44779a02f515c39e712bb7cc4f8a85610efe739f"} Mar 12 18:24:57 crc kubenswrapper[4926]: I0312 18:24:57.436141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"759fd18072cdf8fcc7bc2d92cc950b5720a437d7e4487f5098fffd2244e21cde"} Mar 12 18:24:57 crc kubenswrapper[4926]: I0312 18:24:57.436164 4926 scope.go:117] "RemoveContainer" containerID="10c4816f4e2fc4ce2bc2183a633d9bc53980639515bfce0cf198e862b133fadb" Mar 12 18:24:58 crc kubenswrapper[4926]: I0312 18:24:58.799900 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.375063 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-88dpb"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.377080 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.379392 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.379999 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.386644 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-88dpb"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.559513 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.561242 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.563003 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-scripts\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.563037 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.563084 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxvsd\" (UniqueName: \"kubernetes.io/projected/d30718d9-b986-4490-9e43-56eaa459aeb5-kube-api-access-mxvsd\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.563134 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-config-data\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.572893 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.573043 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.594538 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.595962 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.610823 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.617821 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.655531 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.656613 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.659513 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.668890 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjrtd\" (UniqueName: \"kubernetes.io/projected/522020cf-1556-4192-92c8-6cab42123da0-kube-api-access-hjrtd\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.669010 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-scripts\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.669034 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.669084 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-config-data\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.669111 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxvsd\" (UniqueName: \"kubernetes.io/projected/d30718d9-b986-4490-9e43-56eaa459aeb5-kube-api-access-mxvsd\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.669143 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522020cf-1556-4192-92c8-6cab42123da0-logs\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.669177 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.669209 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-config-data\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.676482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-scripts\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.690008 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.697669 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxvsd\" (UniqueName: \"kubernetes.io/projected/d30718d9-b986-4490-9e43-56eaa459aeb5-kube-api-access-mxvsd\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.698048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-config-data\") pod \"nova-cell0-cell-mapping-88dpb\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.752644 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773500 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4307f4cd-b3d5-41cf-94ba-055b0f09c920-logs\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773583 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m94n\" (UniqueName: \"kubernetes.io/projected/4307f4cd-b3d5-41cf-94ba-055b0f09c920-kube-api-access-9m94n\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773602 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-config-data\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-config-data\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773683 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss276\" (UniqueName: \"kubernetes.io/projected/7c00983f-ac91-410c-9ee3-d55342198d70-kube-api-access-ss276\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773717 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-config-data\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773753 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522020cf-1556-4192-92c8-6cab42123da0-logs\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773776 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773795 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.773822 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjrtd\" (UniqueName: \"kubernetes.io/projected/522020cf-1556-4192-92c8-6cab42123da0-kube-api-access-hjrtd\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.779031 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522020cf-1556-4192-92c8-6cab42123da0-logs\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.785955 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.801119 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-config-data\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.804501 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwx72"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.805952 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.817852 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwx72"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.831498 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.832692 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.840718 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.844703 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.848466 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjrtd\" (UniqueName: \"kubernetes.io/projected/522020cf-1556-4192-92c8-6cab42123da0-kube-api-access-hjrtd\") pod \"nova-api-0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.875827 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m94n\" (UniqueName: \"kubernetes.io/projected/4307f4cd-b3d5-41cf-94ba-055b0f09c920-kube-api-access-9m94n\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.875868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-config-data\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.875934 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-config-data\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.875959 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss276\" (UniqueName: \"kubernetes.io/projected/7c00983f-ac91-410c-9ee3-d55342198d70-kube-api-access-ss276\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.876024 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.876056 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4307f4cd-b3d5-41cf-94ba-055b0f09c920-logs\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.876072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.877961 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4307f4cd-b3d5-41cf-94ba-055b0f09c920-logs\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.881398 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.884814 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-config-data\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.887932 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.893676 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-config-data\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.893716 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.895919 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss276\" (UniqueName: \"kubernetes.io/projected/7c00983f-ac91-410c-9ee3-d55342198d70-kube-api-access-ss276\") pod \"nova-scheduler-0\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.898361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m94n\" (UniqueName: \"kubernetes.io/projected/4307f4cd-b3d5-41cf-94ba-055b0f09c920-kube-api-access-9m94n\") pod \"nova-metadata-0\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.927347 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.977474 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.977529 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.977555 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppkxw\" (UniqueName: \"kubernetes.io/projected/eba9290a-15ad-403b-901b-bab0917192dd-kube-api-access-ppkxw\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.977574 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.977607 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.977866 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.978004 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7xnp\" (UniqueName: \"kubernetes.io/projected/458c7cf7-4e8a-4272-8940-1d730293a0ca-kube-api-access-k7xnp\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.978038 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.978065 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.978141 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-config\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:24:59 crc kubenswrapper[4926]: I0312 18:24:59.997972 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.079917 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-config\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080258 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080308 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080340 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppkxw\" (UniqueName: \"kubernetes.io/projected/eba9290a-15ad-403b-901b-bab0917192dd-kube-api-access-ppkxw\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080402 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080523 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7xnp\" (UniqueName: \"kubernetes.io/projected/458c7cf7-4e8a-4272-8940-1d730293a0ca-kube-api-access-k7xnp\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080557 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.080593 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.081497 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-svc\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.081923 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.083168 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.084384 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.084922 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-config\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.086522 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.086785 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.098017 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppkxw\" (UniqueName: \"kubernetes.io/projected/eba9290a-15ad-403b-901b-bab0917192dd-kube-api-access-ppkxw\") pod \"nova-cell1-novncproxy-0\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.104452 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7xnp\" (UniqueName: \"kubernetes.io/projected/458c7cf7-4e8a-4272-8940-1d730293a0ca-kube-api-access-k7xnp\") pod \"dnsmasq-dns-757b4f8459-jwx72\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.211403 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.215641 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.455376 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.581213 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.714145 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.776503 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-88dpb"] Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.797912 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2wcs"] Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.802488 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.808145 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.808622 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.809467 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2wcs"] Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.899460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-config-data\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.899499 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-scripts\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.899625 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:00 crc kubenswrapper[4926]: I0312 18:25:00.899716 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t75rc\" (UniqueName: \"kubernetes.io/projected/b478ba05-2155-4b13-a58a-002be25403d0-kube-api-access-t75rc\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.001005 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.001154 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t75rc\" (UniqueName: \"kubernetes.io/projected/b478ba05-2155-4b13-a58a-002be25403d0-kube-api-access-t75rc\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.001236 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-config-data\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.001265 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-scripts\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.006667 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.014161 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-scripts\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.026198 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-config-data\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.026668 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t75rc\" (UniqueName: \"kubernetes.io/projected/b478ba05-2155-4b13-a58a-002be25403d0-kube-api-access-t75rc\") pod \"nova-cell1-conductor-db-sync-x2wcs\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.032221 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.051717 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwx72"] Mar 12 18:25:01 crc kubenswrapper[4926]: W0312 18:25:01.052952 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod458c7cf7_4e8a_4272_8940_1d730293a0ca.slice/crio-820539f8ae4b6d013f20317aef2ceed339b3e7e027d5738aa8b884143c185f6f WatchSource:0}: Error finding container 820539f8ae4b6d013f20317aef2ceed339b3e7e027d5738aa8b884143c185f6f: Status 404 returned error can't find the container with id 820539f8ae4b6d013f20317aef2ceed339b3e7e027d5738aa8b884143c185f6f Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.203132 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:01 crc kubenswrapper[4926]: E0312 18:25:01.444777 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified: decoding bearer token (last URL \"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_1b89217552bc42d1be3fb06a1aed001a&scope=repository%3Apodified-antelope-centos9%2Fopenstack-nova-novncproxy%3Apull&service=quay.io\", body start \"\"): unexpected end of JSON input" image="quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified" Mar 12 18:25:01 crc kubenswrapper[4926]: E0312 18:25:01.445332 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell1-novncproxy-novncproxy,Image:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfbhc9h65ch67ch588h57bhd9hb9h98hbbh649h655h67ch55ch689h596h594h668h595h64fh68bh64ch66h696h669h576h56dh58bh59dh54dhffh8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-novncproxy-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ppkxw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/vnc_lite.html,Port:{0 6080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:10,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/vnc_lite.html,Port:{0 6080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/vnc_lite.html,Port:{0 6080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell1-novncproxy-0_openstack(eba9290a-15ad-403b-901b-bab0917192dd): ErrImagePull: initializing source docker://quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified: decoding bearer token (last URL \"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_1b89217552bc42d1be3fb06a1aed001a&scope=repository%3Apodified-antelope-centos9%2Fopenstack-nova-novncproxy%3Apull&service=quay.io\", body start \"\"): unexpected end of JSON input" logger="UnhandledError" Mar 12 18:25:01 crc kubenswrapper[4926]: E0312 18:25:01.446800 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell1-novncproxy-novncproxy\" with ErrImagePull: \"initializing source docker://quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified: decoding bearer token (last URL \\\"https://quay.io/v2/auth?account=openshift-release-dev%2Bocm_access_1b89217552bc42d1be3fb06a1aed001a&scope=repository%3Apodified-antelope-centos9%2Fopenstack-nova-novncproxy%3Apull&service=quay.io\\\", body start \\\"\\\"): unexpected end of JSON input\"" pod="openstack/nova-cell1-novncproxy-0" podUID="eba9290a-15ad-403b-901b-bab0917192dd" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.491888 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"eba9290a-15ad-403b-901b-bab0917192dd","Type":"ContainerStarted","Data":"50f6a42234fb9c45d38c1b23471febadd40f957d027586ffd43709de0e0faeab"} Mar 12 18:25:01 crc kubenswrapper[4926]: E0312 18:25:01.497302 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell1-novncproxy-novncproxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified\\\"\"" pod="openstack/nova-cell1-novncproxy-0" podUID="eba9290a-15ad-403b-901b-bab0917192dd" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.502696 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7c00983f-ac91-410c-9ee3-d55342198d70","Type":"ContainerStarted","Data":"8a2d87ff1b1f13bbcd71a5ed9af411076eaddd0c93ccca3397ca3bb1a0eb3af8"} Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.515764 4926 generic.go:334] "Generic (PLEG): container finished" podID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerID="12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491" exitCode=0 Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.515862 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" event={"ID":"458c7cf7-4e8a-4272-8940-1d730293a0ca","Type":"ContainerDied","Data":"12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491"} Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.515888 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" event={"ID":"458c7cf7-4e8a-4272-8940-1d730293a0ca","Type":"ContainerStarted","Data":"820539f8ae4b6d013f20317aef2ceed339b3e7e027d5738aa8b884143c185f6f"} Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.526887 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4307f4cd-b3d5-41cf-94ba-055b0f09c920","Type":"ContainerStarted","Data":"2f3d265c00f5c2ef87695352db3e88d3c30936c8db618015345f1cf9afeddcc5"} Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.537339 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-88dpb" event={"ID":"d30718d9-b986-4490-9e43-56eaa459aeb5","Type":"ContainerStarted","Data":"2f4d726cd118a123d7b32175fc4e520ff68fb0c877630e02cc5267fde94f47a0"} Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.537381 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-88dpb" event={"ID":"d30718d9-b986-4490-9e43-56eaa459aeb5","Type":"ContainerStarted","Data":"c008177eb27a46bebec75159ab0b2a0dca1d32e65ccd34eeac95828b596395d8"} Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.545955 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522020cf-1556-4192-92c8-6cab42123da0","Type":"ContainerStarted","Data":"42ea86479f09e663945c5ef83197d6d342bd9e261315d58ca9f838ea2ad55e1c"} Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.581867 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-88dpb" podStartSLOduration=2.581834546 podStartE2EDuration="2.581834546s" podCreationTimestamp="2026-03-12 18:24:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:01.557129521 +0000 UTC m=+1341.925755854" watchObservedRunningTime="2026-03-12 18:25:01.581834546 +0000 UTC m=+1341.950460879" Mar 12 18:25:01 crc kubenswrapper[4926]: I0312 18:25:01.716334 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2wcs"] Mar 12 18:25:01 crc kubenswrapper[4926]: W0312 18:25:01.724424 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb478ba05_2155_4b13_a58a_002be25403d0.slice/crio-7299f6360af3c9f86bc4fe3b2e6a46529a369ee63e25d1be1c606c3639d19f01 WatchSource:0}: Error finding container 7299f6360af3c9f86bc4fe3b2e6a46529a369ee63e25d1be1c606c3639d19f01: Status 404 returned error can't find the container with id 7299f6360af3c9f86bc4fe3b2e6a46529a369ee63e25d1be1c606c3639d19f01 Mar 12 18:25:02 crc kubenswrapper[4926]: I0312 18:25:02.565950 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" event={"ID":"b478ba05-2155-4b13-a58a-002be25403d0","Type":"ContainerStarted","Data":"14cbad167526774b9aa42516a514a2cffabd952a439286d5c8bfe698bbd81e8b"} Mar 12 18:25:02 crc kubenswrapper[4926]: I0312 18:25:02.566547 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" event={"ID":"b478ba05-2155-4b13-a58a-002be25403d0","Type":"ContainerStarted","Data":"7299f6360af3c9f86bc4fe3b2e6a46529a369ee63e25d1be1c606c3639d19f01"} Mar 12 18:25:02 crc kubenswrapper[4926]: I0312 18:25:02.570324 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" event={"ID":"458c7cf7-4e8a-4272-8940-1d730293a0ca","Type":"ContainerStarted","Data":"9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3"} Mar 12 18:25:02 crc kubenswrapper[4926]: E0312 18:25:02.571514 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell1-novncproxy-novncproxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified\\\"\"" pod="openstack/nova-cell1-novncproxy-0" podUID="eba9290a-15ad-403b-901b-bab0917192dd" Mar 12 18:25:02 crc kubenswrapper[4926]: I0312 18:25:02.582920 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" podStartSLOduration=2.582902079 podStartE2EDuration="2.582902079s" podCreationTimestamp="2026-03-12 18:25:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:02.580507544 +0000 UTC m=+1342.949133877" watchObservedRunningTime="2026-03-12 18:25:02.582902079 +0000 UTC m=+1342.951528412" Mar 12 18:25:02 crc kubenswrapper[4926]: I0312 18:25:02.597275 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" podStartSLOduration=3.59725317 podStartE2EDuration="3.59725317s" podCreationTimestamp="2026-03-12 18:24:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:02.595895497 +0000 UTC m=+1342.964521830" watchObservedRunningTime="2026-03-12 18:25:02.59725317 +0000 UTC m=+1342.965879503" Mar 12 18:25:03 crc kubenswrapper[4926]: I0312 18:25:03.406459 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:03 crc kubenswrapper[4926]: I0312 18:25:03.447302 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:25:03 crc kubenswrapper[4926]: I0312 18:25:03.580130 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7c00983f-ac91-410c-9ee3-d55342198d70","Type":"ContainerStarted","Data":"ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323"} Mar 12 18:25:03 crc kubenswrapper[4926]: I0312 18:25:03.584842 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4307f4cd-b3d5-41cf-94ba-055b0f09c920","Type":"ContainerStarted","Data":"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91"} Mar 12 18:25:03 crc kubenswrapper[4926]: I0312 18:25:03.587741 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522020cf-1556-4192-92c8-6cab42123da0","Type":"ContainerStarted","Data":"2acf5477ea08c1fb97c59765c73d7abfdadc6189fbd1cf4549d66c1007c1d817"} Mar 12 18:25:03 crc kubenswrapper[4926]: I0312 18:25:03.588805 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:03 crc kubenswrapper[4926]: I0312 18:25:03.613728 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.205749246 podStartE2EDuration="4.613702255s" podCreationTimestamp="2026-03-12 18:24:59 +0000 UTC" firstStartedPulling="2026-03-12 18:25:00.684905672 +0000 UTC m=+1341.053532015" lastFinishedPulling="2026-03-12 18:25:03.092858691 +0000 UTC m=+1343.461485024" observedRunningTime="2026-03-12 18:25:03.598229789 +0000 UTC m=+1343.966856142" watchObservedRunningTime="2026-03-12 18:25:03.613702255 +0000 UTC m=+1343.982328588" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.067912 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.203197 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-config-data\") pod \"eba9290a-15ad-403b-901b-bab0917192dd\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.203285 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppkxw\" (UniqueName: \"kubernetes.io/projected/eba9290a-15ad-403b-901b-bab0917192dd-kube-api-access-ppkxw\") pod \"eba9290a-15ad-403b-901b-bab0917192dd\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.203325 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-combined-ca-bundle\") pod \"eba9290a-15ad-403b-901b-bab0917192dd\" (UID: \"eba9290a-15ad-403b-901b-bab0917192dd\") " Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.209588 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eba9290a-15ad-403b-901b-bab0917192dd-kube-api-access-ppkxw" (OuterVolumeSpecName: "kube-api-access-ppkxw") pod "eba9290a-15ad-403b-901b-bab0917192dd" (UID: "eba9290a-15ad-403b-901b-bab0917192dd"). InnerVolumeSpecName "kube-api-access-ppkxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.210342 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eba9290a-15ad-403b-901b-bab0917192dd" (UID: "eba9290a-15ad-403b-901b-bab0917192dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.222125 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-config-data" (OuterVolumeSpecName: "config-data") pod "eba9290a-15ad-403b-901b-bab0917192dd" (UID: "eba9290a-15ad-403b-901b-bab0917192dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.306030 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.306088 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppkxw\" (UniqueName: \"kubernetes.io/projected/eba9290a-15ad-403b-901b-bab0917192dd-kube-api-access-ppkxw\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.306110 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eba9290a-15ad-403b-901b-bab0917192dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.598838 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4307f4cd-b3d5-41cf-94ba-055b0f09c920","Type":"ContainerStarted","Data":"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69"} Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.598917 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-log" containerID="cri-o://7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91" gracePeriod=30 Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.598955 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-metadata" containerID="cri-o://aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69" gracePeriod=30 Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.603419 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522020cf-1556-4192-92c8-6cab42123da0","Type":"ContainerStarted","Data":"d312e71f700c8194066a4bfb0efbe92c4e2b9fbced6805317853fff15c35d5d6"} Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.609043 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"eba9290a-15ad-403b-901b-bab0917192dd","Type":"ContainerDied","Data":"50f6a42234fb9c45d38c1b23471febadd40f957d027586ffd43709de0e0faeab"} Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.609119 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.625654 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.111126465 podStartE2EDuration="5.625632127s" podCreationTimestamp="2026-03-12 18:24:59 +0000 UTC" firstStartedPulling="2026-03-12 18:25:00.578843944 +0000 UTC m=+1340.947470277" lastFinishedPulling="2026-03-12 18:25:03.093349606 +0000 UTC m=+1343.461975939" observedRunningTime="2026-03-12 18:25:04.621706435 +0000 UTC m=+1344.990332768" watchObservedRunningTime="2026-03-12 18:25:04.625632127 +0000 UTC m=+1344.994258470" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.645615 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.993577968 podStartE2EDuration="5.645596254s" podCreationTimestamp="2026-03-12 18:24:59 +0000 UTC" firstStartedPulling="2026-03-12 18:25:00.467123649 +0000 UTC m=+1340.835749982" lastFinishedPulling="2026-03-12 18:25:03.119141935 +0000 UTC m=+1343.487768268" observedRunningTime="2026-03-12 18:25:04.645092568 +0000 UTC m=+1345.013718911" watchObservedRunningTime="2026-03-12 18:25:04.645596254 +0000 UTC m=+1345.014222597" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.689910 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.700348 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.710178 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.711748 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.718603 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.718821 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.718960 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.719479 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.814759 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.814814 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.814907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhmpz\" (UniqueName: \"kubernetes.io/projected/5990afe5-179c-401d-99eb-58b27e2bfc9e-kube-api-access-lhmpz\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.815028 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.815053 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.916758 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.916816 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.916961 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.916985 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.918141 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhmpz\" (UniqueName: \"kubernetes.io/projected/5990afe5-179c-401d-99eb-58b27e2bfc9e-kube-api-access-lhmpz\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.921971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.924107 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.924905 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.928900 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.928964 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.931179 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5990afe5-179c-401d-99eb-58b27e2bfc9e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.949354 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhmpz\" (UniqueName: \"kubernetes.io/projected/5990afe5-179c-401d-99eb-58b27e2bfc9e-kube-api-access-lhmpz\") pod \"nova-cell1-novncproxy-0\" (UID: \"5990afe5-179c-401d-99eb-58b27e2bfc9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:04 crc kubenswrapper[4926]: I0312 18:25:04.979002 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.077156 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.383324 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.536418 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-config-data\") pod \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.536880 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m94n\" (UniqueName: \"kubernetes.io/projected/4307f4cd-b3d5-41cf-94ba-055b0f09c920-kube-api-access-9m94n\") pod \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.537049 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4307f4cd-b3d5-41cf-94ba-055b0f09c920-logs\") pod \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.537094 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-combined-ca-bundle\") pod \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\" (UID: \"4307f4cd-b3d5-41cf-94ba-055b0f09c920\") " Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.537791 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4307f4cd-b3d5-41cf-94ba-055b0f09c920-logs" (OuterVolumeSpecName: "logs") pod "4307f4cd-b3d5-41cf-94ba-055b0f09c920" (UID: "4307f4cd-b3d5-41cf-94ba-055b0f09c920"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.541649 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4307f4cd-b3d5-41cf-94ba-055b0f09c920-kube-api-access-9m94n" (OuterVolumeSpecName: "kube-api-access-9m94n") pod "4307f4cd-b3d5-41cf-94ba-055b0f09c920" (UID: "4307f4cd-b3d5-41cf-94ba-055b0f09c920"). InnerVolumeSpecName "kube-api-access-9m94n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.569398 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-config-data" (OuterVolumeSpecName: "config-data") pod "4307f4cd-b3d5-41cf-94ba-055b0f09c920" (UID: "4307f4cd-b3d5-41cf-94ba-055b0f09c920"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.581590 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4307f4cd-b3d5-41cf-94ba-055b0f09c920" (UID: "4307f4cd-b3d5-41cf-94ba-055b0f09c920"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.620663 4926 generic.go:334] "Generic (PLEG): container finished" podID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerID="aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69" exitCode=0 Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.620695 4926 generic.go:334] "Generic (PLEG): container finished" podID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerID="7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91" exitCode=143 Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.620728 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.620763 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4307f4cd-b3d5-41cf-94ba-055b0f09c920","Type":"ContainerDied","Data":"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69"} Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.620790 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4307f4cd-b3d5-41cf-94ba-055b0f09c920","Type":"ContainerDied","Data":"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91"} Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.620799 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4307f4cd-b3d5-41cf-94ba-055b0f09c920","Type":"ContainerDied","Data":"2f3d265c00f5c2ef87695352db3e88d3c30936c8db618015345f1cf9afeddcc5"} Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.620844 4926 scope.go:117] "RemoveContainer" containerID="aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.639413 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4307f4cd-b3d5-41cf-94ba-055b0f09c920-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.639526 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.639545 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4307f4cd-b3d5-41cf-94ba-055b0f09c920-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.639563 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m94n\" (UniqueName: \"kubernetes.io/projected/4307f4cd-b3d5-41cf-94ba-055b0f09c920-kube-api-access-9m94n\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.657548 4926 scope.go:117] "RemoveContainer" containerID="7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.675608 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.711033 4926 scope.go:117] "RemoveContainer" containerID="aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69" Mar 12 18:25:05 crc kubenswrapper[4926]: E0312 18:25:05.711541 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69\": container with ID starting with aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69 not found: ID does not exist" containerID="aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.711605 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69"} err="failed to get container status \"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69\": rpc error: code = NotFound desc = could not find container \"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69\": container with ID starting with aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69 not found: ID does not exist" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.711634 4926 scope.go:117] "RemoveContainer" containerID="7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91" Mar 12 18:25:05 crc kubenswrapper[4926]: E0312 18:25:05.712071 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91\": container with ID starting with 7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91 not found: ID does not exist" containerID="7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.712226 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91"} err="failed to get container status \"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91\": rpc error: code = NotFound desc = could not find container \"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91\": container with ID starting with 7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91 not found: ID does not exist" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.712362 4926 scope.go:117] "RemoveContainer" containerID="aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.713016 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69"} err="failed to get container status \"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69\": rpc error: code = NotFound desc = could not find container \"aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69\": container with ID starting with aa88e9c555f20e87aa7f99a35587675471ba4c00aa7e3cf59b886406679b6d69 not found: ID does not exist" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.713086 4926 scope.go:117] "RemoveContainer" containerID="7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.713561 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91"} err="failed to get container status \"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91\": rpc error: code = NotFound desc = could not find container \"7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91\": container with ID starting with 7547facc65c6eefbe2350b5bf9bbb600c1c3ab2aee9b66cde1ae452f75229e91 not found: ID does not exist" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.731297 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.744615 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.753500 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:05 crc kubenswrapper[4926]: E0312 18:25:05.754020 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-metadata" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.754047 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-metadata" Mar 12 18:25:05 crc kubenswrapper[4926]: E0312 18:25:05.754079 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-log" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.754088 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-log" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.754309 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-metadata" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.754327 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" containerName="nova-metadata-log" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.755776 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.757840 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.759929 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.764210 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.945141 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.945207 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-config-data\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.945321 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.945346 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1344f618-3fb1-4718-ac28-aaac70801a43-logs\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:05 crc kubenswrapper[4926]: I0312 18:25:05.945375 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkw7f\" (UniqueName: \"kubernetes.io/projected/1344f618-3fb1-4718-ac28-aaac70801a43-kube-api-access-bkw7f\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.048033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.048118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-config-data\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.048283 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.048327 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1344f618-3fb1-4718-ac28-aaac70801a43-logs\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.048386 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkw7f\" (UniqueName: \"kubernetes.io/projected/1344f618-3fb1-4718-ac28-aaac70801a43-kube-api-access-bkw7f\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.049280 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1344f618-3fb1-4718-ac28-aaac70801a43-logs\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.056192 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-config-data\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.056802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.067614 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.074921 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkw7f\" (UniqueName: \"kubernetes.io/projected/1344f618-3fb1-4718-ac28-aaac70801a43-kube-api-access-bkw7f\") pod \"nova-metadata-0\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.080119 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.502899 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4307f4cd-b3d5-41cf-94ba-055b0f09c920" path="/var/lib/kubelet/pods/4307f4cd-b3d5-41cf-94ba-055b0f09c920/volumes" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.504268 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eba9290a-15ad-403b-901b-bab0917192dd" path="/var/lib/kubelet/pods/eba9290a-15ad-403b-901b-bab0917192dd/volumes" Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.565408 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:06 crc kubenswrapper[4926]: I0312 18:25:06.635189 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5990afe5-179c-401d-99eb-58b27e2bfc9e","Type":"ContainerStarted","Data":"9632a93bdd6e87ebaad7f0644944aca90fb862d669cf288ad2f9ce7ad904f639"} Mar 12 18:25:06 crc kubenswrapper[4926]: W0312 18:25:06.894270 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1344f618_3fb1_4718_ac28_aaac70801a43.slice/crio-21a532f650953a29a6610144508556c5c62c3cfee82633d3ccf1f934e89f4ad7 WatchSource:0}: Error finding container 21a532f650953a29a6610144508556c5c62c3cfee82633d3ccf1f934e89f4ad7: Status 404 returned error can't find the container with id 21a532f650953a29a6610144508556c5c62c3cfee82633d3ccf1f934e89f4ad7 Mar 12 18:25:07 crc kubenswrapper[4926]: I0312 18:25:07.649578 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5990afe5-179c-401d-99eb-58b27e2bfc9e","Type":"ContainerStarted","Data":"8004a0c4872d595eff541c18db2d835ce9f58755458fed592f03016d57654f99"} Mar 12 18:25:07 crc kubenswrapper[4926]: I0312 18:25:07.655428 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1344f618-3fb1-4718-ac28-aaac70801a43","Type":"ContainerStarted","Data":"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db"} Mar 12 18:25:07 crc kubenswrapper[4926]: I0312 18:25:07.655509 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1344f618-3fb1-4718-ac28-aaac70801a43","Type":"ContainerStarted","Data":"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88"} Mar 12 18:25:07 crc kubenswrapper[4926]: I0312 18:25:07.655523 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1344f618-3fb1-4718-ac28-aaac70801a43","Type":"ContainerStarted","Data":"21a532f650953a29a6610144508556c5c62c3cfee82633d3ccf1f934e89f4ad7"} Mar 12 18:25:07 crc kubenswrapper[4926]: I0312 18:25:07.679721 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.374399772 podStartE2EDuration="3.679704611s" podCreationTimestamp="2026-03-12 18:25:04 +0000 UTC" firstStartedPulling="2026-03-12 18:25:05.665424875 +0000 UTC m=+1346.034051198" lastFinishedPulling="2026-03-12 18:25:06.970729704 +0000 UTC m=+1347.339356037" observedRunningTime="2026-03-12 18:25:07.677081198 +0000 UTC m=+1348.045707541" watchObservedRunningTime="2026-03-12 18:25:07.679704611 +0000 UTC m=+1348.048330944" Mar 12 18:25:07 crc kubenswrapper[4926]: I0312 18:25:07.703598 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.70357994 podStartE2EDuration="2.70357994s" podCreationTimestamp="2026-03-12 18:25:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:07.701901317 +0000 UTC m=+1348.070527690" watchObservedRunningTime="2026-03-12 18:25:07.70357994 +0000 UTC m=+1348.072206273" Mar 12 18:25:08 crc kubenswrapper[4926]: I0312 18:25:08.680329 4926 generic.go:334] "Generic (PLEG): container finished" podID="d30718d9-b986-4490-9e43-56eaa459aeb5" containerID="2f4d726cd118a123d7b32175fc4e520ff68fb0c877630e02cc5267fde94f47a0" exitCode=0 Mar 12 18:25:08 crc kubenswrapper[4926]: I0312 18:25:08.680402 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-88dpb" event={"ID":"d30718d9-b986-4490-9e43-56eaa459aeb5","Type":"ContainerDied","Data":"2f4d726cd118a123d7b32175fc4e520ff68fb0c877630e02cc5267fde94f47a0"} Mar 12 18:25:09 crc kubenswrapper[4926]: I0312 18:25:09.699623 4926 generic.go:334] "Generic (PLEG): container finished" podID="b478ba05-2155-4b13-a58a-002be25403d0" containerID="14cbad167526774b9aa42516a514a2cffabd952a439286d5c8bfe698bbd81e8b" exitCode=0 Mar 12 18:25:09 crc kubenswrapper[4926]: I0312 18:25:09.699775 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" event={"ID":"b478ba05-2155-4b13-a58a-002be25403d0","Type":"ContainerDied","Data":"14cbad167526774b9aa42516a514a2cffabd952a439286d5c8bfe698bbd81e8b"} Mar 12 18:25:09 crc kubenswrapper[4926]: I0312 18:25:09.889474 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 18:25:09 crc kubenswrapper[4926]: I0312 18:25:09.889713 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 18:25:09 crc kubenswrapper[4926]: I0312 18:25:09.979268 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.016783 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.077999 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.140853 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.213654 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.233093 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-scripts\") pod \"d30718d9-b986-4490-9e43-56eaa459aeb5\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.233164 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-config-data\") pod \"d30718d9-b986-4490-9e43-56eaa459aeb5\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.233235 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-combined-ca-bundle\") pod \"d30718d9-b986-4490-9e43-56eaa459aeb5\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.233330 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxvsd\" (UniqueName: \"kubernetes.io/projected/d30718d9-b986-4490-9e43-56eaa459aeb5-kube-api-access-mxvsd\") pod \"d30718d9-b986-4490-9e43-56eaa459aeb5\" (UID: \"d30718d9-b986-4490-9e43-56eaa459aeb5\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.251226 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d30718d9-b986-4490-9e43-56eaa459aeb5-kube-api-access-mxvsd" (OuterVolumeSpecName: "kube-api-access-mxvsd") pod "d30718d9-b986-4490-9e43-56eaa459aeb5" (UID: "d30718d9-b986-4490-9e43-56eaa459aeb5"). InnerVolumeSpecName "kube-api-access-mxvsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.284020 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-scripts" (OuterVolumeSpecName: "scripts") pod "d30718d9-b986-4490-9e43-56eaa459aeb5" (UID: "d30718d9-b986-4490-9e43-56eaa459aeb5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.313244 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rln7t"] Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.313964 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" podUID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerName="dnsmasq-dns" containerID="cri-o://ad9f00f4a3b221ab329deab372c49c3816c4c0e0f8a0ad87e7aefa0a45c31c5b" gracePeriod=10 Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.321974 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-config-data" (OuterVolumeSpecName: "config-data") pod "d30718d9-b986-4490-9e43-56eaa459aeb5" (UID: "d30718d9-b986-4490-9e43-56eaa459aeb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.336514 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d30718d9-b986-4490-9e43-56eaa459aeb5" (UID: "d30718d9-b986-4490-9e43-56eaa459aeb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.341198 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.341225 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.341236 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30718d9-b986-4490-9e43-56eaa459aeb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.341248 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxvsd\" (UniqueName: \"kubernetes.io/projected/d30718d9-b986-4490-9e43-56eaa459aeb5-kube-api-access-mxvsd\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: E0312 18:25:10.557985 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd30718d9_b986_4490_9e43_56eaa459aeb5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75ab53a0_e811_4b31_8a1e_d71be4115b2c.slice/crio-ad9f00f4a3b221ab329deab372c49c3816c4c0e0f8a0ad87e7aefa0a45c31c5b.scope\": RecentStats: unable to find data in memory cache]" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.712308 4926 generic.go:334] "Generic (PLEG): container finished" podID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerID="ad9f00f4a3b221ab329deab372c49c3816c4c0e0f8a0ad87e7aefa0a45c31c5b" exitCode=0 Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.712384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" event={"ID":"75ab53a0-e811-4b31-8a1e-d71be4115b2c","Type":"ContainerDied","Data":"ad9f00f4a3b221ab329deab372c49c3816c4c0e0f8a0ad87e7aefa0a45c31c5b"} Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.714265 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-88dpb" event={"ID":"d30718d9-b986-4490-9e43-56eaa459aeb5","Type":"ContainerDied","Data":"c008177eb27a46bebec75159ab0b2a0dca1d32e65ccd34eeac95828b596395d8"} Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.714314 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c008177eb27a46bebec75159ab0b2a0dca1d32e65ccd34eeac95828b596395d8" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.714369 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-88dpb" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.752361 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.765487 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.850784 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-config\") pod \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.850856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-sb\") pod \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.850980 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-swift-storage-0\") pod \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.851011 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-svc\") pod \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.851101 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm8xc\" (UniqueName: \"kubernetes.io/projected/75ab53a0-e811-4b31-8a1e-d71be4115b2c-kube-api-access-sm8xc\") pod \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.851135 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-nb\") pod \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\" (UID: \"75ab53a0-e811-4b31-8a1e-d71be4115b2c\") " Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.855875 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75ab53a0-e811-4b31-8a1e-d71be4115b2c-kube-api-access-sm8xc" (OuterVolumeSpecName: "kube-api-access-sm8xc") pod "75ab53a0-e811-4b31-8a1e-d71be4115b2c" (UID: "75ab53a0-e811-4b31-8a1e-d71be4115b2c"). InnerVolumeSpecName "kube-api-access-sm8xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.912927 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "75ab53a0-e811-4b31-8a1e-d71be4115b2c" (UID: "75ab53a0-e811-4b31-8a1e-d71be4115b2c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.929652 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "75ab53a0-e811-4b31-8a1e-d71be4115b2c" (UID: "75ab53a0-e811-4b31-8a1e-d71be4115b2c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.931978 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-config" (OuterVolumeSpecName: "config") pod "75ab53a0-e811-4b31-8a1e-d71be4115b2c" (UID: "75ab53a0-e811-4b31-8a1e-d71be4115b2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.947747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "75ab53a0-e811-4b31-8a1e-d71be4115b2c" (UID: "75ab53a0-e811-4b31-8a1e-d71be4115b2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.950339 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "75ab53a0-e811-4b31-8a1e-d71be4115b2c" (UID: "75ab53a0-e811-4b31-8a1e-d71be4115b2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.968731 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm8xc\" (UniqueName: \"kubernetes.io/projected/75ab53a0-e811-4b31-8a1e-d71be4115b2c-kube-api-access-sm8xc\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.968765 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.968775 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.968783 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.968792 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:10 crc kubenswrapper[4926]: I0312 18:25:10.968800 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75ab53a0-e811-4b31-8a1e-d71be4115b2c-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.007107 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.007449 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.019552 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.019908 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-log" containerID="cri-o://2acf5477ea08c1fb97c59765c73d7abfdadc6189fbd1cf4549d66c1007c1d817" gracePeriod=30 Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.020328 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-api" containerID="cri-o://d312e71f700c8194066a4bfb0efbe92c4e2b9fbced6805317853fff15c35d5d6" gracePeriod=30 Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.042254 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.042482 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-log" containerID="cri-o://cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88" gracePeriod=30 Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.042553 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-metadata" containerID="cri-o://f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db" gracePeriod=30 Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.080544 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.080616 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.111882 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.273769 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-combined-ca-bundle\") pod \"b478ba05-2155-4b13-a58a-002be25403d0\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.273972 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-scripts\") pod \"b478ba05-2155-4b13-a58a-002be25403d0\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.274461 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t75rc\" (UniqueName: \"kubernetes.io/projected/b478ba05-2155-4b13-a58a-002be25403d0-kube-api-access-t75rc\") pod \"b478ba05-2155-4b13-a58a-002be25403d0\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.274488 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-config-data\") pod \"b478ba05-2155-4b13-a58a-002be25403d0\" (UID: \"b478ba05-2155-4b13-a58a-002be25403d0\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.288590 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.289203 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-scripts" (OuterVolumeSpecName: "scripts") pod "b478ba05-2155-4b13-a58a-002be25403d0" (UID: "b478ba05-2155-4b13-a58a-002be25403d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.289284 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b478ba05-2155-4b13-a58a-002be25403d0-kube-api-access-t75rc" (OuterVolumeSpecName: "kube-api-access-t75rc") pod "b478ba05-2155-4b13-a58a-002be25403d0" (UID: "b478ba05-2155-4b13-a58a-002be25403d0"). InnerVolumeSpecName "kube-api-access-t75rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.309200 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b478ba05-2155-4b13-a58a-002be25403d0" (UID: "b478ba05-2155-4b13-a58a-002be25403d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.322648 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-config-data" (OuterVolumeSpecName: "config-data") pod "b478ba05-2155-4b13-a58a-002be25403d0" (UID: "b478ba05-2155-4b13-a58a-002be25403d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.377958 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.378212 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t75rc\" (UniqueName: \"kubernetes.io/projected/b478ba05-2155-4b13-a58a-002be25403d0-kube-api-access-t75rc\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.378225 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.378236 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b478ba05-2155-4b13-a58a-002be25403d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.567868 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.687542 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-combined-ca-bundle\") pod \"1344f618-3fb1-4718-ac28-aaac70801a43\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.687685 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-config-data\") pod \"1344f618-3fb1-4718-ac28-aaac70801a43\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.687905 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1344f618-3fb1-4718-ac28-aaac70801a43-logs\") pod \"1344f618-3fb1-4718-ac28-aaac70801a43\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.688363 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1344f618-3fb1-4718-ac28-aaac70801a43-logs" (OuterVolumeSpecName: "logs") pod "1344f618-3fb1-4718-ac28-aaac70801a43" (UID: "1344f618-3fb1-4718-ac28-aaac70801a43"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.688430 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-nova-metadata-tls-certs\") pod \"1344f618-3fb1-4718-ac28-aaac70801a43\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.688704 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkw7f\" (UniqueName: \"kubernetes.io/projected/1344f618-3fb1-4718-ac28-aaac70801a43-kube-api-access-bkw7f\") pod \"1344f618-3fb1-4718-ac28-aaac70801a43\" (UID: \"1344f618-3fb1-4718-ac28-aaac70801a43\") " Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.689840 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1344f618-3fb1-4718-ac28-aaac70801a43-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.691316 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1344f618-3fb1-4718-ac28-aaac70801a43-kube-api-access-bkw7f" (OuterVolumeSpecName: "kube-api-access-bkw7f") pod "1344f618-3fb1-4718-ac28-aaac70801a43" (UID: "1344f618-3fb1-4718-ac28-aaac70801a43"). InnerVolumeSpecName "kube-api-access-bkw7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.717527 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-config-data" (OuterVolumeSpecName: "config-data") pod "1344f618-3fb1-4718-ac28-aaac70801a43" (UID: "1344f618-3fb1-4718-ac28-aaac70801a43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.722332 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1344f618-3fb1-4718-ac28-aaac70801a43" (UID: "1344f618-3fb1-4718-ac28-aaac70801a43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.724814 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.724810 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-x2wcs" event={"ID":"b478ba05-2155-4b13-a58a-002be25403d0","Type":"ContainerDied","Data":"7299f6360af3c9f86bc4fe3b2e6a46529a369ee63e25d1be1c606c3639d19f01"} Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.724924 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7299f6360af3c9f86bc4fe3b2e6a46529a369ee63e25d1be1c606c3639d19f01" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.727323 4926 generic.go:334] "Generic (PLEG): container finished" podID="1344f618-3fb1-4718-ac28-aaac70801a43" containerID="f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db" exitCode=0 Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.727350 4926 generic.go:334] "Generic (PLEG): container finished" podID="1344f618-3fb1-4718-ac28-aaac70801a43" containerID="cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88" exitCode=143 Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.727396 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1344f618-3fb1-4718-ac28-aaac70801a43","Type":"ContainerDied","Data":"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db"} Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.727418 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1344f618-3fb1-4718-ac28-aaac70801a43","Type":"ContainerDied","Data":"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88"} Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.727431 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1344f618-3fb1-4718-ac28-aaac70801a43","Type":"ContainerDied","Data":"21a532f650953a29a6610144508556c5c62c3cfee82633d3ccf1f934e89f4ad7"} Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.727469 4926 scope.go:117] "RemoveContainer" containerID="f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.727604 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.732933 4926 generic.go:334] "Generic (PLEG): container finished" podID="522020cf-1556-4192-92c8-6cab42123da0" containerID="2acf5477ea08c1fb97c59765c73d7abfdadc6189fbd1cf4549d66c1007c1d817" exitCode=143 Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.733001 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522020cf-1556-4192-92c8-6cab42123da0","Type":"ContainerDied","Data":"2acf5477ea08c1fb97c59765c73d7abfdadc6189fbd1cf4549d66c1007c1d817"} Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.739752 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1344f618-3fb1-4718-ac28-aaac70801a43" (UID: "1344f618-3fb1-4718-ac28-aaac70801a43"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.756092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" event={"ID":"75ab53a0-e811-4b31-8a1e-d71be4115b2c","Type":"ContainerDied","Data":"2c7e6fa4e701c31af5cac5d50047d5137f377c93ee8e669dc7d858e4f08b2574"} Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.756119 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-rln7t" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.773141 4926 scope.go:117] "RemoveContainer" containerID="cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.792756 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.792783 4926 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.792795 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkw7f\" (UniqueName: \"kubernetes.io/projected/1344f618-3fb1-4718-ac28-aaac70801a43-kube-api-access-bkw7f\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.792803 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1344f618-3fb1-4718-ac28-aaac70801a43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.803698 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rln7t"] Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.813664 4926 scope.go:117] "RemoveContainer" containerID="f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.814824 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db\": container with ID starting with f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db not found: ID does not exist" containerID="f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.814876 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db"} err="failed to get container status \"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db\": rpc error: code = NotFound desc = could not find container \"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db\": container with ID starting with f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db not found: ID does not exist" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.814897 4926 scope.go:117] "RemoveContainer" containerID="cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.816765 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88\": container with ID starting with cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88 not found: ID does not exist" containerID="cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.816800 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88"} err="failed to get container status \"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88\": rpc error: code = NotFound desc = could not find container \"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88\": container with ID starting with cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88 not found: ID does not exist" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.816818 4926 scope.go:117] "RemoveContainer" containerID="f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.817071 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db"} err="failed to get container status \"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db\": rpc error: code = NotFound desc = could not find container \"f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db\": container with ID starting with f1295472ac240b745f8f6e8144f3cf3a94d1eb3adaa6cb76a9497a26654594db not found: ID does not exist" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.817123 4926 scope.go:117] "RemoveContainer" containerID="cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.817417 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88"} err="failed to get container status \"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88\": rpc error: code = NotFound desc = could not find container \"cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88\": container with ID starting with cc38b398f5f07a27852a773cc63f78a5c3c6858e17355598c8a7599e76f24a88 not found: ID does not exist" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.817537 4926 scope.go:117] "RemoveContainer" containerID="ad9f00f4a3b221ab329deab372c49c3816c4c0e0f8a0ad87e7aefa0a45c31c5b" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.820903 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-rln7t"] Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.842950 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.843418 4926 scope.go:117] "RemoveContainer" containerID="ba45c947e1da77d586b46aef916a37581f794d00be599bae5b1f6a31185bfcd3" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.843783 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerName="init" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.843802 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerName="init" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.843816 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerName="dnsmasq-dns" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.843822 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerName="dnsmasq-dns" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.843837 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-metadata" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.843843 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-metadata" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.843855 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b478ba05-2155-4b13-a58a-002be25403d0" containerName="nova-cell1-conductor-db-sync" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.843860 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b478ba05-2155-4b13-a58a-002be25403d0" containerName="nova-cell1-conductor-db-sync" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.843873 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-log" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.843878 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-log" Mar 12 18:25:11 crc kubenswrapper[4926]: E0312 18:25:11.843897 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30718d9-b986-4490-9e43-56eaa459aeb5" containerName="nova-manage" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.843902 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30718d9-b986-4490-9e43-56eaa459aeb5" containerName="nova-manage" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.844077 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30718d9-b986-4490-9e43-56eaa459aeb5" containerName="nova-manage" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.844087 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-log" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.844094 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" containerName="dnsmasq-dns" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.844110 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b478ba05-2155-4b13-a58a-002be25403d0" containerName="nova-cell1-conductor-db-sync" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.844117 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" containerName="nova-metadata-metadata" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.846034 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.850865 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.851343 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.894667 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r866l\" (UniqueName: \"kubernetes.io/projected/9696ed5e-263b-457b-a771-9f7a7e27dbed-kube-api-access-r866l\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.894728 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9696ed5e-263b-457b-a771-9f7a7e27dbed-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.894753 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9696ed5e-263b-457b-a771-9f7a7e27dbed-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.996853 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r866l\" (UniqueName: \"kubernetes.io/projected/9696ed5e-263b-457b-a771-9f7a7e27dbed-kube-api-access-r866l\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.996914 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9696ed5e-263b-457b-a771-9f7a7e27dbed-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:11 crc kubenswrapper[4926]: I0312 18:25:11.996934 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9696ed5e-263b-457b-a771-9f7a7e27dbed-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.000674 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9696ed5e-263b-457b-a771-9f7a7e27dbed-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.001425 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9696ed5e-263b-457b-a771-9f7a7e27dbed-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.016362 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r866l\" (UniqueName: \"kubernetes.io/projected/9696ed5e-263b-457b-a771-9f7a7e27dbed-kube-api-access-r866l\") pod \"nova-cell1-conductor-0\" (UID: \"9696ed5e-263b-457b-a771-9f7a7e27dbed\") " pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.140930 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.152919 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.169779 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.171807 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.172328 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.174386 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.174661 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.180934 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.301881 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4zmx\" (UniqueName: \"kubernetes.io/projected/f6b42e45-6f11-4207-b26d-7befa423860f-kube-api-access-b4zmx\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.301957 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b42e45-6f11-4207-b26d-7befa423860f-logs\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.301987 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.302056 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-config-data\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.302134 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.404288 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b42e45-6f11-4207-b26d-7befa423860f-logs\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.404344 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.404409 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-config-data\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.404490 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.404557 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4zmx\" (UniqueName: \"kubernetes.io/projected/f6b42e45-6f11-4207-b26d-7befa423860f-kube-api-access-b4zmx\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.405410 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b42e45-6f11-4207-b26d-7befa423860f-logs\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.411701 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-config-data\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.412006 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.417913 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.425941 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4zmx\" (UniqueName: \"kubernetes.io/projected/f6b42e45-6f11-4207-b26d-7befa423860f-kube-api-access-b4zmx\") pod \"nova-metadata-0\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.505544 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1344f618-3fb1-4718-ac28-aaac70801a43" path="/var/lib/kubelet/pods/1344f618-3fb1-4718-ac28-aaac70801a43/volumes" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.507002 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75ab53a0-e811-4b31-8a1e-d71be4115b2c" path="/var/lib/kubelet/pods/75ab53a0-e811-4b31-8a1e-d71be4115b2c/volumes" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.586638 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.660559 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 12 18:25:12 crc kubenswrapper[4926]: W0312 18:25:12.662220 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9696ed5e_263b_457b_a771_9f7a7e27dbed.slice/crio-ac5d920764343c3822aea3eec4f980699931e85fcb37936f0c6999814073f8f3 WatchSource:0}: Error finding container ac5d920764343c3822aea3eec4f980699931e85fcb37936f0c6999814073f8f3: Status 404 returned error can't find the container with id ac5d920764343c3822aea3eec4f980699931e85fcb37936f0c6999814073f8f3 Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.801231 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7c00983f-ac91-410c-9ee3-d55342198d70" containerName="nova-scheduler-scheduler" containerID="cri-o://ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" gracePeriod=30 Mar 12 18:25:12 crc kubenswrapper[4926]: I0312 18:25:12.801563 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9696ed5e-263b-457b-a771-9f7a7e27dbed","Type":"ContainerStarted","Data":"ac5d920764343c3822aea3eec4f980699931e85fcb37936f0c6999814073f8f3"} Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.081907 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.812412 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"9696ed5e-263b-457b-a771-9f7a7e27dbed","Type":"ContainerStarted","Data":"bb9e08c6749891ff5759e1975e0508d47c8fe6e0741407ca15c1c009eb4de943"} Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.813771 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.819132 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b42e45-6f11-4207-b26d-7befa423860f","Type":"ContainerStarted","Data":"fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6"} Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.819268 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b42e45-6f11-4207-b26d-7befa423860f","Type":"ContainerStarted","Data":"b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28"} Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.819353 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b42e45-6f11-4207-b26d-7befa423860f","Type":"ContainerStarted","Data":"0161e0bff07323bc954df2ec0573f151d9db197c8f01a596b3666178a9a120fe"} Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.831859 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.8318353480000003 podStartE2EDuration="2.831835348s" podCreationTimestamp="2026-03-12 18:25:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:13.827416899 +0000 UTC m=+1354.196043232" watchObservedRunningTime="2026-03-12 18:25:13.831835348 +0000 UTC m=+1354.200461681" Mar 12 18:25:13 crc kubenswrapper[4926]: I0312 18:25:13.854359 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.8543362540000001 podStartE2EDuration="1.854336254s" podCreationTimestamp="2026-03-12 18:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:13.846390544 +0000 UTC m=+1354.215016897" watchObservedRunningTime="2026-03-12 18:25:13.854336254 +0000 UTC m=+1354.222962587" Mar 12 18:25:14 crc kubenswrapper[4926]: E0312 18:25:14.980380 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323 is running failed: container process not found" containerID="ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 18:25:14 crc kubenswrapper[4926]: E0312 18:25:14.981558 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323 is running failed: container process not found" containerID="ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 18:25:14 crc kubenswrapper[4926]: E0312 18:25:14.981915 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323 is running failed: container process not found" containerID="ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 12 18:25:14 crc kubenswrapper[4926]: E0312 18:25:14.982053 4926 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7c00983f-ac91-410c-9ee3-d55342198d70" containerName="nova-scheduler-scheduler" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.078155 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.096914 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.429637 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.475279 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-combined-ca-bundle\") pod \"7c00983f-ac91-410c-9ee3-d55342198d70\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.475384 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-config-data\") pod \"7c00983f-ac91-410c-9ee3-d55342198d70\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.475597 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss276\" (UniqueName: \"kubernetes.io/projected/7c00983f-ac91-410c-9ee3-d55342198d70-kube-api-access-ss276\") pod \"7c00983f-ac91-410c-9ee3-d55342198d70\" (UID: \"7c00983f-ac91-410c-9ee3-d55342198d70\") " Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.484757 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c00983f-ac91-410c-9ee3-d55342198d70-kube-api-access-ss276" (OuterVolumeSpecName: "kube-api-access-ss276") pod "7c00983f-ac91-410c-9ee3-d55342198d70" (UID: "7c00983f-ac91-410c-9ee3-d55342198d70"). InnerVolumeSpecName "kube-api-access-ss276". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.523182 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-config-data" (OuterVolumeSpecName: "config-data") pod "7c00983f-ac91-410c-9ee3-d55342198d70" (UID: "7c00983f-ac91-410c-9ee3-d55342198d70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.527101 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7c00983f-ac91-410c-9ee3-d55342198d70" (UID: "7c00983f-ac91-410c-9ee3-d55342198d70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.578655 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss276\" (UniqueName: \"kubernetes.io/projected/7c00983f-ac91-410c-9ee3-d55342198d70-kube-api-access-ss276\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.578686 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.578700 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c00983f-ac91-410c-9ee3-d55342198d70-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.841746 4926 generic.go:334] "Generic (PLEG): container finished" podID="7c00983f-ac91-410c-9ee3-d55342198d70" containerID="ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" exitCode=0 Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.841807 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.841820 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7c00983f-ac91-410c-9ee3-d55342198d70","Type":"ContainerDied","Data":"ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323"} Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.841881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7c00983f-ac91-410c-9ee3-d55342198d70","Type":"ContainerDied","Data":"8a2d87ff1b1f13bbcd71a5ed9af411076eaddd0c93ccca3397ca3bb1a0eb3af8"} Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.841903 4926 scope.go:117] "RemoveContainer" containerID="ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.884115 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.884265 4926 scope.go:117] "RemoveContainer" containerID="ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.885349 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 12 18:25:15 crc kubenswrapper[4926]: E0312 18:25:15.893750 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323\": container with ID starting with ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323 not found: ID does not exist" containerID="ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.893793 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323"} err="failed to get container status \"ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323\": rpc error: code = NotFound desc = could not find container \"ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323\": container with ID starting with ea5ca88836de9cfdc6fc5b152d149a335a9b476f1550663f5769f0a8fc968323 not found: ID does not exist" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.914142 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.926394 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:15 crc kubenswrapper[4926]: E0312 18:25:15.926849 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c00983f-ac91-410c-9ee3-d55342198d70" containerName="nova-scheduler-scheduler" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.926869 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c00983f-ac91-410c-9ee3-d55342198d70" containerName="nova-scheduler-scheduler" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.927121 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c00983f-ac91-410c-9ee3-d55342198d70" containerName="nova-scheduler-scheduler" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.927822 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.929945 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.941606 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.987089 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xl7g\" (UniqueName: \"kubernetes.io/projected/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-kube-api-access-4xl7g\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.987243 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-config-data\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:15 crc kubenswrapper[4926]: I0312 18:25:15.987304 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.089786 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-config-data\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.089856 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.089928 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xl7g\" (UniqueName: \"kubernetes.io/projected/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-kube-api-access-4xl7g\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.093828 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.094081 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-config-data\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.104327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xl7g\" (UniqueName: \"kubernetes.io/projected/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-kube-api-access-4xl7g\") pod \"nova-scheduler-0\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.253297 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.502087 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c00983f-ac91-410c-9ee3-d55342198d70" path="/var/lib/kubelet/pods/7c00983f-ac91-410c-9ee3-d55342198d70/volumes" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.753180 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:16 crc kubenswrapper[4926]: W0312 18:25:16.764713 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6e2c8ff_6e0f_4a2b_a6d5_e2c4cf9c97a3.slice/crio-7a7d155b546588f16ab2cdd1af883265350417efed1c7aef091311c1c7d02134 WatchSource:0}: Error finding container 7a7d155b546588f16ab2cdd1af883265350417efed1c7aef091311c1c7d02134: Status 404 returned error can't find the container with id 7a7d155b546588f16ab2cdd1af883265350417efed1c7aef091311c1c7d02134 Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.854485 4926 generic.go:334] "Generic (PLEG): container finished" podID="522020cf-1556-4192-92c8-6cab42123da0" containerID="d312e71f700c8194066a4bfb0efbe92c4e2b9fbced6805317853fff15c35d5d6" exitCode=0 Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.854602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522020cf-1556-4192-92c8-6cab42123da0","Type":"ContainerDied","Data":"d312e71f700c8194066a4bfb0efbe92c4e2b9fbced6805317853fff15c35d5d6"} Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.854631 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"522020cf-1556-4192-92c8-6cab42123da0","Type":"ContainerDied","Data":"42ea86479f09e663945c5ef83197d6d342bd9e261315d58ca9f838ea2ad55e1c"} Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.854644 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42ea86479f09e663945c5ef83197d6d342bd9e261315d58ca9f838ea2ad55e1c" Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.856360 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3","Type":"ContainerStarted","Data":"7a7d155b546588f16ab2cdd1af883265350417efed1c7aef091311c1c7d02134"} Mar 12 18:25:16 crc kubenswrapper[4926]: I0312 18:25:16.908851 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.007718 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjrtd\" (UniqueName: \"kubernetes.io/projected/522020cf-1556-4192-92c8-6cab42123da0-kube-api-access-hjrtd\") pod \"522020cf-1556-4192-92c8-6cab42123da0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.007844 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-combined-ca-bundle\") pod \"522020cf-1556-4192-92c8-6cab42123da0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.007906 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-config-data\") pod \"522020cf-1556-4192-92c8-6cab42123da0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.007959 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522020cf-1556-4192-92c8-6cab42123da0-logs\") pod \"522020cf-1556-4192-92c8-6cab42123da0\" (UID: \"522020cf-1556-4192-92c8-6cab42123da0\") " Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.009753 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/522020cf-1556-4192-92c8-6cab42123da0-logs" (OuterVolumeSpecName: "logs") pod "522020cf-1556-4192-92c8-6cab42123da0" (UID: "522020cf-1556-4192-92c8-6cab42123da0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.011407 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/522020cf-1556-4192-92c8-6cab42123da0-kube-api-access-hjrtd" (OuterVolumeSpecName: "kube-api-access-hjrtd") pod "522020cf-1556-4192-92c8-6cab42123da0" (UID: "522020cf-1556-4192-92c8-6cab42123da0"). InnerVolumeSpecName "kube-api-access-hjrtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.032984 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "522020cf-1556-4192-92c8-6cab42123da0" (UID: "522020cf-1556-4192-92c8-6cab42123da0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.052200 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-config-data" (OuterVolumeSpecName: "config-data") pod "522020cf-1556-4192-92c8-6cab42123da0" (UID: "522020cf-1556-4192-92c8-6cab42123da0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.110630 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.110668 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/522020cf-1556-4192-92c8-6cab42123da0-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.110680 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/522020cf-1556-4192-92c8-6cab42123da0-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.110692 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjrtd\" (UniqueName: \"kubernetes.io/projected/522020cf-1556-4192-92c8-6cab42123da0-kube-api-access-hjrtd\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.208985 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.587706 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.587780 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.725134 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.796987 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-vwzff"] Mar 12 18:25:17 crc kubenswrapper[4926]: E0312 18:25:17.797587 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-api" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.797615 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-api" Mar 12 18:25:17 crc kubenswrapper[4926]: E0312 18:25:17.797638 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-log" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.797649 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-log" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.797969 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-log" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.798013 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="522020cf-1556-4192-92c8-6cab42123da0" containerName="nova-api-api" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.798763 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.801869 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.802416 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.810936 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vwzff"] Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.868582 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.872525 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3","Type":"ContainerStarted","Data":"1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177"} Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.898800 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8987822640000003 podStartE2EDuration="2.898782264s" podCreationTimestamp="2026-03-12 18:25:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:17.889614736 +0000 UTC m=+1358.258241079" watchObservedRunningTime="2026-03-12 18:25:17.898782264 +0000 UTC m=+1358.267408597" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.937724 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrrql\" (UniqueName: \"kubernetes.io/projected/76a36d0b-fd08-4c93-846c-688c71055113-kube-api-access-nrrql\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.937792 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-config-data\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.938027 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-scripts\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.938127 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.939635 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.963564 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.980082 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.983039 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.986279 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 18:25:17 crc kubenswrapper[4926]: I0312 18:25:17.991269 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040576 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5mxb\" (UniqueName: \"kubernetes.io/projected/ff10ffb2-f421-4a3a-aafd-5b081df5d109-kube-api-access-r5mxb\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040626 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrrql\" (UniqueName: \"kubernetes.io/projected/76a36d0b-fd08-4c93-846c-688c71055113-kube-api-access-nrrql\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-config-data\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff10ffb2-f421-4a3a-aafd-5b081df5d109-logs\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040719 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-config-data\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040734 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040779 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-scripts\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.040800 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.046904 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-config-data\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.048009 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-scripts\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.049756 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.058107 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrrql\" (UniqueName: \"kubernetes.io/projected/76a36d0b-fd08-4c93-846c-688c71055113-kube-api-access-nrrql\") pod \"nova-cell1-cell-mapping-vwzff\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.125509 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.142972 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff10ffb2-f421-4a3a-aafd-5b081df5d109-logs\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.143326 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-config-data\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.143342 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.143456 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff10ffb2-f421-4a3a-aafd-5b081df5d109-logs\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.143490 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5mxb\" (UniqueName: \"kubernetes.io/projected/ff10ffb2-f421-4a3a-aafd-5b081df5d109-kube-api-access-r5mxb\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.148936 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.149840 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-config-data\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.161626 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5mxb\" (UniqueName: \"kubernetes.io/projected/ff10ffb2-f421-4a3a-aafd-5b081df5d109-kube-api-access-r5mxb\") pod \"nova-api-0\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.309153 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.534105 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="522020cf-1556-4192-92c8-6cab42123da0" path="/var/lib/kubelet/pods/522020cf-1556-4192-92c8-6cab42123da0/volumes" Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.673898 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-vwzff"] Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.868994 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:18 crc kubenswrapper[4926]: W0312 18:25:18.875385 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff10ffb2_f421_4a3a_aafd_5b081df5d109.slice/crio-daa31bd5c5f69662abb20c63efc93566505de27536b89db47ab73ffecd04929a WatchSource:0}: Error finding container daa31bd5c5f69662abb20c63efc93566505de27536b89db47ab73ffecd04929a: Status 404 returned error can't find the container with id daa31bd5c5f69662abb20c63efc93566505de27536b89db47ab73ffecd04929a Mar 12 18:25:18 crc kubenswrapper[4926]: I0312 18:25:18.877559 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vwzff" event={"ID":"76a36d0b-fd08-4c93-846c-688c71055113","Type":"ContainerStarted","Data":"a7717c4f2567107bbbfa3a1e3937a74f1ee5564e6be712ecb32bfbf5d5d798bb"} Mar 12 18:25:19 crc kubenswrapper[4926]: I0312 18:25:19.889719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vwzff" event={"ID":"76a36d0b-fd08-4c93-846c-688c71055113","Type":"ContainerStarted","Data":"ddcfcec936ed5e6887603a245e014b4f9336991c54bce882c0b7657f1d5a318b"} Mar 12 18:25:19 crc kubenswrapper[4926]: I0312 18:25:19.892782 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff10ffb2-f421-4a3a-aafd-5b081df5d109","Type":"ContainerStarted","Data":"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8"} Mar 12 18:25:19 crc kubenswrapper[4926]: I0312 18:25:19.892859 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff10ffb2-f421-4a3a-aafd-5b081df5d109","Type":"ContainerStarted","Data":"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728"} Mar 12 18:25:19 crc kubenswrapper[4926]: I0312 18:25:19.892878 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff10ffb2-f421-4a3a-aafd-5b081df5d109","Type":"ContainerStarted","Data":"daa31bd5c5f69662abb20c63efc93566505de27536b89db47ab73ffecd04929a"} Mar 12 18:25:19 crc kubenswrapper[4926]: I0312 18:25:19.919485 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-vwzff" podStartSLOduration=2.9194579000000003 podStartE2EDuration="2.9194579s" podCreationTimestamp="2026-03-12 18:25:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:19.91528595 +0000 UTC m=+1360.283912303" watchObservedRunningTime="2026-03-12 18:25:19.9194579 +0000 UTC m=+1360.288084243" Mar 12 18:25:19 crc kubenswrapper[4926]: I0312 18:25:19.942423 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.94240352 podStartE2EDuration="2.94240352s" podCreationTimestamp="2026-03-12 18:25:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:19.936678451 +0000 UTC m=+1360.305304784" watchObservedRunningTime="2026-03-12 18:25:19.94240352 +0000 UTC m=+1360.311029853" Mar 12 18:25:21 crc kubenswrapper[4926]: I0312 18:25:21.253948 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 12 18:25:22 crc kubenswrapper[4926]: I0312 18:25:22.552047 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:25:22 crc kubenswrapper[4926]: I0312 18:25:22.552492 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3343d19e-07d3-4de8-954a-f7e31aa8279f" containerName="kube-state-metrics" containerID="cri-o://c936fa8962854c1d8665ba026930a8b325915da3444b99cba8eeb7debd2bb042" gracePeriod=30 Mar 12 18:25:22 crc kubenswrapper[4926]: I0312 18:25:22.588037 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 18:25:22 crc kubenswrapper[4926]: I0312 18:25:22.588128 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 18:25:22 crc kubenswrapper[4926]: I0312 18:25:22.702023 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="3343d19e-07d3-4de8-954a-f7e31aa8279f" containerName="kube-state-metrics" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": dial tcp 10.217.0.106:8081: connect: connection refused" Mar 12 18:25:22 crc kubenswrapper[4926]: I0312 18:25:22.927105 4926 generic.go:334] "Generic (PLEG): container finished" podID="3343d19e-07d3-4de8-954a-f7e31aa8279f" containerID="c936fa8962854c1d8665ba026930a8b325915da3444b99cba8eeb7debd2bb042" exitCode=2 Mar 12 18:25:22 crc kubenswrapper[4926]: I0312 18:25:22.927425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3343d19e-07d3-4de8-954a-f7e31aa8279f","Type":"ContainerDied","Data":"c936fa8962854c1d8665ba026930a8b325915da3444b99cba8eeb7debd2bb042"} Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.047080 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.143835 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bjbq\" (UniqueName: \"kubernetes.io/projected/3343d19e-07d3-4de8-954a-f7e31aa8279f-kube-api-access-6bjbq\") pod \"3343d19e-07d3-4de8-954a-f7e31aa8279f\" (UID: \"3343d19e-07d3-4de8-954a-f7e31aa8279f\") " Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.164751 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3343d19e-07d3-4de8-954a-f7e31aa8279f-kube-api-access-6bjbq" (OuterVolumeSpecName: "kube-api-access-6bjbq") pod "3343d19e-07d3-4de8-954a-f7e31aa8279f" (UID: "3343d19e-07d3-4de8-954a-f7e31aa8279f"). InnerVolumeSpecName "kube-api-access-6bjbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.245716 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bjbq\" (UniqueName: \"kubernetes.io/projected/3343d19e-07d3-4de8-954a-f7e31aa8279f-kube-api-access-6bjbq\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.601693 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.601709 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.952140 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.952155 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3343d19e-07d3-4de8-954a-f7e31aa8279f","Type":"ContainerDied","Data":"a5a3fb3c5e3c756132a6d36939e2dc50de833c191b34382d6e423a84d04ecf45"} Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.952505 4926 scope.go:117] "RemoveContainer" containerID="c936fa8962854c1d8665ba026930a8b325915da3444b99cba8eeb7debd2bb042" Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.956781 4926 generic.go:334] "Generic (PLEG): container finished" podID="76a36d0b-fd08-4c93-846c-688c71055113" containerID="ddcfcec936ed5e6887603a245e014b4f9336991c54bce882c0b7657f1d5a318b" exitCode=0 Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.956834 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vwzff" event={"ID":"76a36d0b-fd08-4c93-846c-688c71055113","Type":"ContainerDied","Data":"ddcfcec936ed5e6887603a245e014b4f9336991c54bce882c0b7657f1d5a318b"} Mar 12 18:25:23 crc kubenswrapper[4926]: I0312 18:25:23.998647 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.013055 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.025937 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:25:24 crc kubenswrapper[4926]: E0312 18:25:24.026422 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3343d19e-07d3-4de8-954a-f7e31aa8279f" containerName="kube-state-metrics" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.026462 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3343d19e-07d3-4de8-954a-f7e31aa8279f" containerName="kube-state-metrics" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.026804 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3343d19e-07d3-4de8-954a-f7e31aa8279f" containerName="kube-state-metrics" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.027755 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.033375 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.033858 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.035509 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.059713 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.060001 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcclx\" (UniqueName: \"kubernetes.io/projected/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-api-access-dcclx\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.060163 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.060351 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.162124 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.162747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcclx\" (UniqueName: \"kubernetes.io/projected/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-api-access-dcclx\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.162925 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.163101 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.167859 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.168588 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.170267 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.197240 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcclx\" (UniqueName: \"kubernetes.io/projected/30b7fb7c-fcab-4551-8284-c0dab53beb21-kube-api-access-dcclx\") pod \"kube-state-metrics-0\" (UID: \"30b7fb7c-fcab-4551-8284-c0dab53beb21\") " pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.375774 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.471921 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.472328 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="proxy-httpd" containerID="cri-o://271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0" gracePeriod=30 Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.472351 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="sg-core" containerID="cri-o://26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b" gracePeriod=30 Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.472381 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-central-agent" containerID="cri-o://5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d" gracePeriod=30 Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.472353 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-notification-agent" containerID="cri-o://5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d" gracePeriod=30 Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.512372 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3343d19e-07d3-4de8-954a-f7e31aa8279f" path="/var/lib/kubelet/pods/3343d19e-07d3-4de8-954a-f7e31aa8279f/volumes" Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.896977 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 12 18:25:24 crc kubenswrapper[4926]: W0312 18:25:24.912893 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30b7fb7c_fcab_4551_8284_c0dab53beb21.slice/crio-64ca1401b7b6b47ce2ad73b6c528c5b95d0557b055354e78739d6cfd6228b3af WatchSource:0}: Error finding container 64ca1401b7b6b47ce2ad73b6c528c5b95d0557b055354e78739d6cfd6228b3af: Status 404 returned error can't find the container with id 64ca1401b7b6b47ce2ad73b6c528c5b95d0557b055354e78739d6cfd6228b3af Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.966376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30b7fb7c-fcab-4551-8284-c0dab53beb21","Type":"ContainerStarted","Data":"64ca1401b7b6b47ce2ad73b6c528c5b95d0557b055354e78739d6cfd6228b3af"} Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.970779 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerID="271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0" exitCode=0 Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.970803 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerID="26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b" exitCode=2 Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.970811 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerID="5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d" exitCode=0 Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.970829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerDied","Data":"271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0"} Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.970878 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerDied","Data":"26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b"} Mar 12 18:25:24 crc kubenswrapper[4926]: I0312 18:25:24.970891 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerDied","Data":"5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d"} Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.475596 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.592466 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-config-data\") pod \"76a36d0b-fd08-4c93-846c-688c71055113\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.592574 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-scripts\") pod \"76a36d0b-fd08-4c93-846c-688c71055113\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.592629 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-combined-ca-bundle\") pod \"76a36d0b-fd08-4c93-846c-688c71055113\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.592699 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrrql\" (UniqueName: \"kubernetes.io/projected/76a36d0b-fd08-4c93-846c-688c71055113-kube-api-access-nrrql\") pod \"76a36d0b-fd08-4c93-846c-688c71055113\" (UID: \"76a36d0b-fd08-4c93-846c-688c71055113\") " Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.596598 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-scripts" (OuterVolumeSpecName: "scripts") pod "76a36d0b-fd08-4c93-846c-688c71055113" (UID: "76a36d0b-fd08-4c93-846c-688c71055113"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.597272 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a36d0b-fd08-4c93-846c-688c71055113-kube-api-access-nrrql" (OuterVolumeSpecName: "kube-api-access-nrrql") pod "76a36d0b-fd08-4c93-846c-688c71055113" (UID: "76a36d0b-fd08-4c93-846c-688c71055113"). InnerVolumeSpecName "kube-api-access-nrrql". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.623692 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-config-data" (OuterVolumeSpecName: "config-data") pod "76a36d0b-fd08-4c93-846c-688c71055113" (UID: "76a36d0b-fd08-4c93-846c-688c71055113"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.626713 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76a36d0b-fd08-4c93-846c-688c71055113" (UID: "76a36d0b-fd08-4c93-846c-688c71055113"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.694848 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.694881 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.694890 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76a36d0b-fd08-4c93-846c-688c71055113-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.694901 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrrql\" (UniqueName: \"kubernetes.io/projected/76a36d0b-fd08-4c93-846c-688c71055113-kube-api-access-nrrql\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.982270 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-vwzff" event={"ID":"76a36d0b-fd08-4c93-846c-688c71055113","Type":"ContainerDied","Data":"a7717c4f2567107bbbfa3a1e3937a74f1ee5564e6be712ecb32bfbf5d5d798bb"} Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.982311 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7717c4f2567107bbbfa3a1e3937a74f1ee5564e6be712ecb32bfbf5d5d798bb" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.982520 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-vwzff" Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.984353 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"30b7fb7c-fcab-4551-8284-c0dab53beb21","Type":"ContainerStarted","Data":"5a30176409a7d48bc579f1ad02423c28ddc46894468c7dbc8bf6f21ccb4c8522"} Mar 12 18:25:25 crc kubenswrapper[4926]: I0312 18:25:25.984490 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.021363 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.614543336 podStartE2EDuration="3.021343721s" podCreationTimestamp="2026-03-12 18:25:23 +0000 UTC" firstStartedPulling="2026-03-12 18:25:24.916228443 +0000 UTC m=+1365.284854776" lastFinishedPulling="2026-03-12 18:25:25.323028828 +0000 UTC m=+1365.691655161" observedRunningTime="2026-03-12 18:25:26.005275056 +0000 UTC m=+1366.373901389" watchObservedRunningTime="2026-03-12 18:25:26.021343721 +0000 UTC m=+1366.389970064" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.166585 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.167131 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-log" containerID="cri-o://29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728" gracePeriod=30 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.167167 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-api" containerID="cri-o://d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8" gracePeriod=30 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.190566 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.190835 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" containerName="nova-scheduler-scheduler" containerID="cri-o://1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177" gracePeriod=30 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.241034 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.248157 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-log" containerID="cri-o://b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28" gracePeriod=30 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.248333 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-metadata" containerID="cri-o://fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6" gracePeriod=30 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.714604 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.830379 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-combined-ca-bundle\") pod \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.830508 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5mxb\" (UniqueName: \"kubernetes.io/projected/ff10ffb2-f421-4a3a-aafd-5b081df5d109-kube-api-access-r5mxb\") pod \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.830675 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-config-data\") pod \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.830709 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff10ffb2-f421-4a3a-aafd-5b081df5d109-logs\") pod \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\" (UID: \"ff10ffb2-f421-4a3a-aafd-5b081df5d109\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.833607 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff10ffb2-f421-4a3a-aafd-5b081df5d109-logs" (OuterVolumeSpecName: "logs") pod "ff10ffb2-f421-4a3a-aafd-5b081df5d109" (UID: "ff10ffb2-f421-4a3a-aafd-5b081df5d109"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.839653 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff10ffb2-f421-4a3a-aafd-5b081df5d109-kube-api-access-r5mxb" (OuterVolumeSpecName: "kube-api-access-r5mxb") pod "ff10ffb2-f421-4a3a-aafd-5b081df5d109" (UID: "ff10ffb2-f421-4a3a-aafd-5b081df5d109"). InnerVolumeSpecName "kube-api-access-r5mxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.845781 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.859833 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff10ffb2-f421-4a3a-aafd-5b081df5d109" (UID: "ff10ffb2-f421-4a3a-aafd-5b081df5d109"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.861012 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-config-data" (OuterVolumeSpecName: "config-data") pod "ff10ffb2-f421-4a3a-aafd-5b081df5d109" (UID: "ff10ffb2-f421-4a3a-aafd-5b081df5d109"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932130 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-combined-ca-bundle\") pod \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932239 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-log-httpd\") pod \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932313 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-config-data\") pod \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932424 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbct5\" (UniqueName: \"kubernetes.io/projected/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-kube-api-access-dbct5\") pod \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932486 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-scripts\") pod \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932522 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-sg-core-conf-yaml\") pod \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932562 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-run-httpd\") pod \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\" (UID: \"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d\") " Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.932819 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" (UID: "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.933053 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" (UID: "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.933737 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.933760 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff10ffb2-f421-4a3a-aafd-5b081df5d109-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.933769 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff10ffb2-f421-4a3a-aafd-5b081df5d109-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.933779 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.933792 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5mxb\" (UniqueName: \"kubernetes.io/projected/ff10ffb2-f421-4a3a-aafd-5b081df5d109-kube-api-access-r5mxb\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.933802 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.937043 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-scripts" (OuterVolumeSpecName: "scripts") pod "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" (UID: "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.937114 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-kube-api-access-dbct5" (OuterVolumeSpecName: "kube-api-access-dbct5") pod "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" (UID: "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d"). InnerVolumeSpecName "kube-api-access-dbct5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.980144 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" (UID: "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.994295 4926 generic.go:334] "Generic (PLEG): container finished" podID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerID="5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d" exitCode=0 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.994350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerDied","Data":"5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d"} Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.994374 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"6d80b1d3-3647-44fe-9a41-f1a6c11aff1d","Type":"ContainerDied","Data":"ab0418481e92c6856771c7ea5c23c736e2ec58cc43ca849b59c1411888deb2c2"} Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.994391 4926 scope.go:117] "RemoveContainer" containerID="271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.994571 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.999359 4926 generic.go:334] "Generic (PLEG): container finished" podID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerID="d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8" exitCode=0 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.999396 4926 generic.go:334] "Generic (PLEG): container finished" podID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerID="29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728" exitCode=143 Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.999447 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.999473 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff10ffb2-f421-4a3a-aafd-5b081df5d109","Type":"ContainerDied","Data":"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8"} Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.999500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff10ffb2-f421-4a3a-aafd-5b081df5d109","Type":"ContainerDied","Data":"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728"} Mar 12 18:25:26 crc kubenswrapper[4926]: I0312 18:25:26.999509 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff10ffb2-f421-4a3a-aafd-5b081df5d109","Type":"ContainerDied","Data":"daa31bd5c5f69662abb20c63efc93566505de27536b89db47ab73ffecd04929a"} Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.001388 4926 generic.go:334] "Generic (PLEG): container finished" podID="f6b42e45-6f11-4207-b26d-7befa423860f" containerID="b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28" exitCode=143 Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.001470 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b42e45-6f11-4207-b26d-7befa423860f","Type":"ContainerDied","Data":"b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28"} Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.036174 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbct5\" (UniqueName: \"kubernetes.io/projected/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-kube-api-access-dbct5\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.036205 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.036214 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.040504 4926 scope.go:117] "RemoveContainer" containerID="26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.042157 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.054501 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.055714 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" (UID: "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.062867 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-config-data" (OuterVolumeSpecName: "config-data") pod "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" (UID: "6d80b1d3-3647-44fe-9a41-f1a6c11aff1d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.070569 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.071076 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-log" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.071098 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-log" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.071119 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-central-agent" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.071127 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-central-agent" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.071142 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-api" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.071149 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-api" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.071164 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="proxy-httpd" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.071171 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="proxy-httpd" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.071188 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-notification-agent" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.071295 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-notification-agent" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.071318 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="sg-core" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.071325 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="sg-core" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.071338 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a36d0b-fd08-4c93-846c-688c71055113" containerName="nova-manage" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.071345 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a36d0b-fd08-4c93-846c-688c71055113" containerName="nova-manage" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.072271 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a36d0b-fd08-4c93-846c-688c71055113" containerName="nova-manage" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.072294 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="proxy-httpd" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.072314 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-central-agent" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.072325 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="ceilometer-notification-agent" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.072337 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-api" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.072351 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" containerName="nova-api-log" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.072362 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" containerName="sg-core" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.074655 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.078005 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.080009 4926 scope.go:117] "RemoveContainer" containerID="5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.101899 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.112501 4926 scope.go:117] "RemoveContainer" containerID="5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.137548 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-config-data\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.137594 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.137619 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-logs\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.137690 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsrb4\" (UniqueName: \"kubernetes.io/projected/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-kube-api-access-qsrb4\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.137733 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.137744 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.137731 4926 scope.go:117] "RemoveContainer" containerID="271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.138212 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0\": container with ID starting with 271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0 not found: ID does not exist" containerID="271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.138273 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0"} err="failed to get container status \"271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0\": rpc error: code = NotFound desc = could not find container \"271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0\": container with ID starting with 271aa5ed045ae8594c9cbb880d3c37153215115447cba035c80a461aca09cea0 not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.138292 4926 scope.go:117] "RemoveContainer" containerID="26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.138575 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b\": container with ID starting with 26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b not found: ID does not exist" containerID="26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.138608 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b"} err="failed to get container status \"26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b\": rpc error: code = NotFound desc = could not find container \"26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b\": container with ID starting with 26aaa0ccefcdb9eb3bdf054f4109a5704267b244c20325cae1cd1d7a964dc43b not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.138629 4926 scope.go:117] "RemoveContainer" containerID="5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.138833 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d\": container with ID starting with 5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d not found: ID does not exist" containerID="5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.138862 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d"} err="failed to get container status \"5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d\": rpc error: code = NotFound desc = could not find container \"5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d\": container with ID starting with 5133e83018172bddb2651e87d19c3c584b880343960493ab0fec8d492083c66d not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.138878 4926 scope.go:117] "RemoveContainer" containerID="5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.139065 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d\": container with ID starting with 5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d not found: ID does not exist" containerID="5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.139088 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d"} err="failed to get container status \"5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d\": rpc error: code = NotFound desc = could not find container \"5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d\": container with ID starting with 5a7b4815509ad4aad6640591fd715ed5b59dfabe9c09852413d0e4c6fe85885d not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.139151 4926 scope.go:117] "RemoveContainer" containerID="d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.155520 4926 scope.go:117] "RemoveContainer" containerID="29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.185559 4926 scope.go:117] "RemoveContainer" containerID="d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.185996 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8\": container with ID starting with d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8 not found: ID does not exist" containerID="d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.186019 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8"} err="failed to get container status \"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8\": rpc error: code = NotFound desc = could not find container \"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8\": container with ID starting with d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8 not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.186038 4926 scope.go:117] "RemoveContainer" containerID="29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728" Mar 12 18:25:27 crc kubenswrapper[4926]: E0312 18:25:27.186370 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728\": container with ID starting with 29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728 not found: ID does not exist" containerID="29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.186385 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728"} err="failed to get container status \"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728\": rpc error: code = NotFound desc = could not find container \"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728\": container with ID starting with 29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728 not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.186399 4926 scope.go:117] "RemoveContainer" containerID="d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.186612 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8"} err="failed to get container status \"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8\": rpc error: code = NotFound desc = could not find container \"d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8\": container with ID starting with d65cf4c963918ec2d4ea410d0b00104c86b16edaf457c2b52b271c52720055c8 not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.186649 4926 scope.go:117] "RemoveContainer" containerID="29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.186961 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728"} err="failed to get container status \"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728\": rpc error: code = NotFound desc = could not find container \"29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728\": container with ID starting with 29b49f9130da1335dccc5243b155c7585a61ed1e763af8bc4996db2dc9990728 not found: ID does not exist" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.239701 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-config-data\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.239752 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.239780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-logs\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.240128 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsrb4\" (UniqueName: \"kubernetes.io/projected/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-kube-api-access-qsrb4\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.240624 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-logs\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.243813 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-config-data\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.244265 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.256650 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsrb4\" (UniqueName: \"kubernetes.io/projected/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-kube-api-access-qsrb4\") pod \"nova-api-0\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.357780 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.368021 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.379083 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.381708 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.384296 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.385236 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.394168 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.406606 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.411386 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.453178 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.453237 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.453330 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-scripts\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.453355 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.453382 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.453516 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-config-data\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.454824 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh6lh\" (UniqueName: \"kubernetes.io/projected/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-kube-api-access-hh6lh\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.454884 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.557887 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.558627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-scripts\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.558663 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.558686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.558741 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-config-data\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.558795 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh6lh\" (UniqueName: \"kubernetes.io/projected/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-kube-api-access-hh6lh\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.558826 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.558866 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.559862 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.560289 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.564360 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.565213 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-config-data\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.565692 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-scripts\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.566243 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.569006 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.578152 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh6lh\" (UniqueName: \"kubernetes.io/projected/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-kube-api-access-hh6lh\") pod \"ceilometer-0\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.706507 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:27 crc kubenswrapper[4926]: I0312 18:25:27.890136 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:27 crc kubenswrapper[4926]: W0312 18:25:27.896284 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd37e2fa1_9044_4dfb_9aea_bcb441cc91a4.slice/crio-e038299ff9160ecc8a030d680499ae9a217c62acd0df47033360941b4347d3a1 WatchSource:0}: Error finding container e038299ff9160ecc8a030d680499ae9a217c62acd0df47033360941b4347d3a1: Status 404 returned error can't find the container with id e038299ff9160ecc8a030d680499ae9a217c62acd0df47033360941b4347d3a1 Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.020260 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4","Type":"ContainerStarted","Data":"e038299ff9160ecc8a030d680499ae9a217c62acd0df47033360941b4347d3a1"} Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.164579 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:28 crc kubenswrapper[4926]: W0312 18:25:28.165310 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccd8b5fe_8294_40c2_a6f4_3b82012686ec.slice/crio-da79f6ead62afe8afb991cfd28d34557304f0f537134c344d4948f2cb4bfd386 WatchSource:0}: Error finding container da79f6ead62afe8afb991cfd28d34557304f0f537134c344d4948f2cb4bfd386: Status 404 returned error can't find the container with id da79f6ead62afe8afb991cfd28d34557304f0f537134c344d4948f2cb4bfd386 Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.504369 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d80b1d3-3647-44fe-9a41-f1a6c11aff1d" path="/var/lib/kubelet/pods/6d80b1d3-3647-44fe-9a41-f1a6c11aff1d/volumes" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.505799 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff10ffb2-f421-4a3a-aafd-5b081df5d109" path="/var/lib/kubelet/pods/ff10ffb2-f421-4a3a-aafd-5b081df5d109/volumes" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.765569 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.885759 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xl7g\" (UniqueName: \"kubernetes.io/projected/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-kube-api-access-4xl7g\") pod \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.885874 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-combined-ca-bundle\") pod \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.886251 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-config-data\") pod \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\" (UID: \"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3\") " Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.890342 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-kube-api-access-4xl7g" (OuterVolumeSpecName: "kube-api-access-4xl7g") pod "e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" (UID: "e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3"). InnerVolumeSpecName "kube-api-access-4xl7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.917882 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-config-data" (OuterVolumeSpecName: "config-data") pod "e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" (UID: "e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.920800 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" (UID: "e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.988752 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.988802 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:28 crc kubenswrapper[4926]: I0312 18:25:28.988818 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xl7g\" (UniqueName: \"kubernetes.io/projected/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3-kube-api-access-4xl7g\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.030362 4926 generic.go:334] "Generic (PLEG): container finished" podID="e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" containerID="1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177" exitCode=0 Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.030471 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3","Type":"ContainerDied","Data":"1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177"} Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.030510 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3","Type":"ContainerDied","Data":"7a7d155b546588f16ab2cdd1af883265350417efed1c7aef091311c1c7d02134"} Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.030537 4926 scope.go:117] "RemoveContainer" containerID="1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.030709 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.035992 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerStarted","Data":"00d08df89cba0281fd578a80d34eeb606178fe179a18e0bb4dca41d34fcaae6c"} Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.036042 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerStarted","Data":"da79f6ead62afe8afb991cfd28d34557304f0f537134c344d4948f2cb4bfd386"} Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.038170 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4","Type":"ContainerStarted","Data":"6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c"} Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.038202 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4","Type":"ContainerStarted","Data":"a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12"} Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.060667 4926 scope.go:117] "RemoveContainer" containerID="1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177" Mar 12 18:25:29 crc kubenswrapper[4926]: E0312 18:25:29.061097 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177\": container with ID starting with 1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177 not found: ID does not exist" containerID="1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.061132 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177"} err="failed to get container status \"1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177\": rpc error: code = NotFound desc = could not find container \"1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177\": container with ID starting with 1c8ef3a70608e2a8423d2b9d25faab813e85066f6ede6071e20748773e377177 not found: ID does not exist" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.067914 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.067893098 podStartE2EDuration="2.067893098s" podCreationTimestamp="2026-03-12 18:25:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:29.054454606 +0000 UTC m=+1369.423080969" watchObservedRunningTime="2026-03-12 18:25:29.067893098 +0000 UTC m=+1369.436519431" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.094213 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.130426 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.149158 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:29 crc kubenswrapper[4926]: E0312 18:25:29.149631 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" containerName="nova-scheduler-scheduler" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.149654 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" containerName="nova-scheduler-scheduler" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.149854 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" containerName="nova-scheduler-scheduler" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.150456 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.152265 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.160951 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.192963 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9c2983-b118-4677-ba18-20531d4223ad-config-data\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.193053 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9c2983-b118-4677-ba18-20531d4223ad-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.193129 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/6d9c2983-b118-4677-ba18-20531d4223ad-kube-api-access-ffrfd\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.294663 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/6d9c2983-b118-4677-ba18-20531d4223ad-kube-api-access-ffrfd\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.294752 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9c2983-b118-4677-ba18-20531d4223ad-config-data\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.294820 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9c2983-b118-4677-ba18-20531d4223ad-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.299172 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d9c2983-b118-4677-ba18-20531d4223ad-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.304032 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d9c2983-b118-4677-ba18-20531d4223ad-config-data\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.312286 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/6d9c2983-b118-4677-ba18-20531d4223ad-kube-api-access-ffrfd\") pod \"nova-scheduler-0\" (UID: \"6d9c2983-b118-4677-ba18-20531d4223ad\") " pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.520270 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.819295 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.903714 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-combined-ca-bundle\") pod \"f6b42e45-6f11-4207-b26d-7befa423860f\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.903793 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4zmx\" (UniqueName: \"kubernetes.io/projected/f6b42e45-6f11-4207-b26d-7befa423860f-kube-api-access-b4zmx\") pod \"f6b42e45-6f11-4207-b26d-7befa423860f\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.903871 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-config-data\") pod \"f6b42e45-6f11-4207-b26d-7befa423860f\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.903912 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b42e45-6f11-4207-b26d-7befa423860f-logs\") pod \"f6b42e45-6f11-4207-b26d-7befa423860f\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.903964 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-nova-metadata-tls-certs\") pod \"f6b42e45-6f11-4207-b26d-7befa423860f\" (UID: \"f6b42e45-6f11-4207-b26d-7befa423860f\") " Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.909301 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b42e45-6f11-4207-b26d-7befa423860f-kube-api-access-b4zmx" (OuterVolumeSpecName: "kube-api-access-b4zmx") pod "f6b42e45-6f11-4207-b26d-7befa423860f" (UID: "f6b42e45-6f11-4207-b26d-7befa423860f"). InnerVolumeSpecName "kube-api-access-b4zmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.909777 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6b42e45-6f11-4207-b26d-7befa423860f-logs" (OuterVolumeSpecName: "logs") pod "f6b42e45-6f11-4207-b26d-7befa423860f" (UID: "f6b42e45-6f11-4207-b26d-7befa423860f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.930094 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6b42e45-6f11-4207-b26d-7befa423860f" (UID: "f6b42e45-6f11-4207-b26d-7befa423860f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.930708 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-config-data" (OuterVolumeSpecName: "config-data") pod "f6b42e45-6f11-4207-b26d-7befa423860f" (UID: "f6b42e45-6f11-4207-b26d-7befa423860f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:29 crc kubenswrapper[4926]: I0312 18:25:29.946848 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f6b42e45-6f11-4207-b26d-7befa423860f" (UID: "f6b42e45-6f11-4207-b26d-7befa423860f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.006422 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.006466 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4zmx\" (UniqueName: \"kubernetes.io/projected/f6b42e45-6f11-4207-b26d-7befa423860f-kube-api-access-b4zmx\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.006476 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.006485 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f6b42e45-6f11-4207-b26d-7befa423860f-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.006494 4926 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6b42e45-6f11-4207-b26d-7befa423860f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.037784 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.047695 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerStarted","Data":"22a2d1e8587226f797d924ae278b1c0e7b846e0cc05d9f335cbb31ac04af937c"} Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.052267 4926 generic.go:334] "Generic (PLEG): container finished" podID="f6b42e45-6f11-4207-b26d-7befa423860f" containerID="fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6" exitCode=0 Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.052339 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b42e45-6f11-4207-b26d-7befa423860f","Type":"ContainerDied","Data":"fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6"} Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.052371 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f6b42e45-6f11-4207-b26d-7befa423860f","Type":"ContainerDied","Data":"0161e0bff07323bc954df2ec0573f151d9db197c8f01a596b3666178a9a120fe"} Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.052394 4926 scope.go:117] "RemoveContainer" containerID="fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.052575 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.098632 4926 scope.go:117] "RemoveContainer" containerID="b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.112053 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.126283 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.138279 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:30 crc kubenswrapper[4926]: E0312 18:25:30.139185 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-log" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.139214 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-log" Mar 12 18:25:30 crc kubenswrapper[4926]: E0312 18:25:30.139267 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-metadata" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.139277 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-metadata" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.139736 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-log" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.139774 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" containerName="nova-metadata-metadata" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.146970 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.150901 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.151126 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.180491 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.198322 4926 scope.go:117] "RemoveContainer" containerID="fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6" Mar 12 18:25:30 crc kubenswrapper[4926]: E0312 18:25:30.199222 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6\": container with ID starting with fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6 not found: ID does not exist" containerID="fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.199271 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6"} err="failed to get container status \"fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6\": rpc error: code = NotFound desc = could not find container \"fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6\": container with ID starting with fb487b53c7739924b7925ad02cd2780113be299a6f5b88e0215775a40ccaf8b6 not found: ID does not exist" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.199297 4926 scope.go:117] "RemoveContainer" containerID="b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28" Mar 12 18:25:30 crc kubenswrapper[4926]: E0312 18:25:30.199617 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28\": container with ID starting with b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28 not found: ID does not exist" containerID="b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.199643 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28"} err="failed to get container status \"b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28\": rpc error: code = NotFound desc = could not find container \"b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28\": container with ID starting with b0f9e215a663c32f91df3902f5d9ab111e489e136457cd89dc0cefc519b52f28 not found: ID does not exist" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.209392 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8065d406-f127-4c57-b603-e7e6afeb3731-logs\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.209571 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.209761 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.209822 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnbck\" (UniqueName: \"kubernetes.io/projected/8065d406-f127-4c57-b603-e7e6afeb3731-kube-api-access-rnbck\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.209853 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-config-data\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.310926 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.311040 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.311081 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnbck\" (UniqueName: \"kubernetes.io/projected/8065d406-f127-4c57-b603-e7e6afeb3731-kube-api-access-rnbck\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.311108 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-config-data\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.311165 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8065d406-f127-4c57-b603-e7e6afeb3731-logs\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.311571 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8065d406-f127-4c57-b603-e7e6afeb3731-logs\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.319632 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.319939 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.320143 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8065d406-f127-4c57-b603-e7e6afeb3731-config-data\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.338459 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnbck\" (UniqueName: \"kubernetes.io/projected/8065d406-f127-4c57-b603-e7e6afeb3731-kube-api-access-rnbck\") pod \"nova-metadata-0\" (UID: \"8065d406-f127-4c57-b603-e7e6afeb3731\") " pod="openstack/nova-metadata-0" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.507191 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3" path="/var/lib/kubelet/pods/e6e2c8ff-6e0f-4a2b-a6d5-e2c4cf9c97a3/volumes" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.507882 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b42e45-6f11-4207-b26d-7befa423860f" path="/var/lib/kubelet/pods/f6b42e45-6f11-4207-b26d-7befa423860f/volumes" Mar 12 18:25:30 crc kubenswrapper[4926]: I0312 18:25:30.587197 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 12 18:25:31 crc kubenswrapper[4926]: I0312 18:25:31.071148 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d9c2983-b118-4677-ba18-20531d4223ad","Type":"ContainerStarted","Data":"baa7c017e61a4be9bf8c8ad4d676a5457b6ae3959676b8e4e40e02afb89888e2"} Mar 12 18:25:31 crc kubenswrapper[4926]: I0312 18:25:31.071758 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6d9c2983-b118-4677-ba18-20531d4223ad","Type":"ContainerStarted","Data":"4be6a5c77109442ca23bc3eb5c55dcc298e0a416b77d98df56f9405885ecfee6"} Mar 12 18:25:31 crc kubenswrapper[4926]: I0312 18:25:31.075482 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerStarted","Data":"e6bde295c7812497b3848b236f4e958ec3015f8ebeca1a4fdbd6ee1413805cb9"} Mar 12 18:25:31 crc kubenswrapper[4926]: I0312 18:25:31.117176 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.117158521 podStartE2EDuration="2.117158521s" podCreationTimestamp="2026-03-12 18:25:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:31.088993947 +0000 UTC m=+1371.457620290" watchObservedRunningTime="2026-03-12 18:25:31.117158521 +0000 UTC m=+1371.485784854" Mar 12 18:25:31 crc kubenswrapper[4926]: W0312 18:25:31.126581 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8065d406_f127_4c57_b603_e7e6afeb3731.slice/crio-eb3eb7fbcaf5e170aebb941bb0070f18800fda026662721aebf121b775e6664c WatchSource:0}: Error finding container eb3eb7fbcaf5e170aebb941bb0070f18800fda026662721aebf121b775e6664c: Status 404 returned error can't find the container with id eb3eb7fbcaf5e170aebb941bb0070f18800fda026662721aebf121b775e6664c Mar 12 18:25:31 crc kubenswrapper[4926]: I0312 18:25:31.133261 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 12 18:25:32 crc kubenswrapper[4926]: I0312 18:25:32.087887 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8065d406-f127-4c57-b603-e7e6afeb3731","Type":"ContainerStarted","Data":"1e13c7538c29aaae5324b5fbacc1cda6402f336dc0c05bf17ccebd26e5306b6d"} Mar 12 18:25:32 crc kubenswrapper[4926]: I0312 18:25:32.088252 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8065d406-f127-4c57-b603-e7e6afeb3731","Type":"ContainerStarted","Data":"7ef67231a2e8d527454e768f65df4f1e0eaa49d1bdd3fbdd9a31c26ce2bc8392"} Mar 12 18:25:32 crc kubenswrapper[4926]: I0312 18:25:32.088272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8065d406-f127-4c57-b603-e7e6afeb3731","Type":"ContainerStarted","Data":"eb3eb7fbcaf5e170aebb941bb0070f18800fda026662721aebf121b775e6664c"} Mar 12 18:25:32 crc kubenswrapper[4926]: I0312 18:25:32.110589 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.110573024 podStartE2EDuration="2.110573024s" podCreationTimestamp="2026-03-12 18:25:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:32.10725809 +0000 UTC m=+1372.475884443" watchObservedRunningTime="2026-03-12 18:25:32.110573024 +0000 UTC m=+1372.479199357" Mar 12 18:25:34 crc kubenswrapper[4926]: I0312 18:25:34.107110 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerStarted","Data":"2d6e21accb5b365373d4e446af9a0d2f3bb681b5f74d1a1d1d5bbeb21b98d345"} Mar 12 18:25:34 crc kubenswrapper[4926]: I0312 18:25:34.107959 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:25:34 crc kubenswrapper[4926]: I0312 18:25:34.142464 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.863804424 podStartE2EDuration="7.142421601s" podCreationTimestamp="2026-03-12 18:25:27 +0000 UTC" firstStartedPulling="2026-03-12 18:25:28.167914057 +0000 UTC m=+1368.536540390" lastFinishedPulling="2026-03-12 18:25:33.446531234 +0000 UTC m=+1373.815157567" observedRunningTime="2026-03-12 18:25:34.133531852 +0000 UTC m=+1374.502158215" watchObservedRunningTime="2026-03-12 18:25:34.142421601 +0000 UTC m=+1374.511047934" Mar 12 18:25:34 crc kubenswrapper[4926]: I0312 18:25:34.387399 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 12 18:25:34 crc kubenswrapper[4926]: I0312 18:25:34.521636 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 12 18:25:35 crc kubenswrapper[4926]: I0312 18:25:35.588118 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:35 crc kubenswrapper[4926]: I0312 18:25:35.588199 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 12 18:25:37 crc kubenswrapper[4926]: I0312 18:25:37.409490 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 18:25:37 crc kubenswrapper[4926]: I0312 18:25:37.409829 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 18:25:38 crc kubenswrapper[4926]: I0312 18:25:38.498476 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:38 crc kubenswrapper[4926]: I0312 18:25:38.500032 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.209:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:39 crc kubenswrapper[4926]: I0312 18:25:39.521108 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 12 18:25:39 crc kubenswrapper[4926]: I0312 18:25:39.558850 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 12 18:25:40 crc kubenswrapper[4926]: I0312 18:25:40.214807 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 12 18:25:40 crc kubenswrapper[4926]: I0312 18:25:40.588412 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 18:25:40 crc kubenswrapper[4926]: I0312 18:25:40.589894 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 12 18:25:41 crc kubenswrapper[4926]: I0312 18:25:41.606644 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8065d406-f127-4c57-b603-e7e6afeb3731" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:41 crc kubenswrapper[4926]: I0312 18:25:41.606675 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8065d406-f127-4c57-b603-e7e6afeb3731" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 18:25:47 crc kubenswrapper[4926]: I0312 18:25:47.415006 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 18:25:47 crc kubenswrapper[4926]: I0312 18:25:47.416165 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 18:25:47 crc kubenswrapper[4926]: I0312 18:25:47.417200 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 18:25:47 crc kubenswrapper[4926]: I0312 18:25:47.427797 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.282713 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.288226 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.525515 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-nfj99"] Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.527069 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.537186 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-nfj99"] Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.603314 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.603370 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swntz\" (UniqueName: \"kubernetes.io/projected/99d97cdd-8bee-43b1-a07c-fee61fceff3a-kube-api-access-swntz\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.603514 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-config\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.603531 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.603604 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.603706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.705855 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.705909 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swntz\" (UniqueName: \"kubernetes.io/projected/99d97cdd-8bee-43b1-a07c-fee61fceff3a-kube-api-access-swntz\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.705953 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-config\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.705971 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.706005 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.706042 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.706963 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.707370 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.707410 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-config\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.707529 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.707805 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.724038 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swntz\" (UniqueName: \"kubernetes.io/projected/99d97cdd-8bee-43b1-a07c-fee61fceff3a-kube-api-access-swntz\") pod \"dnsmasq-dns-89c5cd4d5-nfj99\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:48 crc kubenswrapper[4926]: I0312 18:25:48.858680 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:49 crc kubenswrapper[4926]: I0312 18:25:49.354553 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-nfj99"] Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.299346 4926 generic.go:334] "Generic (PLEG): container finished" podID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerID="30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0" exitCode=0 Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.300137 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" event={"ID":"99d97cdd-8bee-43b1-a07c-fee61fceff3a","Type":"ContainerDied","Data":"30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0"} Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.300185 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" event={"ID":"99d97cdd-8bee-43b1-a07c-fee61fceff3a","Type":"ContainerStarted","Data":"9711fea2cdf454dba0540ad56623bf62d55c2c749c2e71e1fe9a2ccbb9165424"} Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.472552 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.472819 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-central-agent" containerID="cri-o://00d08df89cba0281fd578a80d34eeb606178fe179a18e0bb4dca41d34fcaae6c" gracePeriod=30 Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.472953 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-notification-agent" containerID="cri-o://22a2d1e8587226f797d924ae278b1c0e7b846e0cc05d9f335cbb31ac04af937c" gracePeriod=30 Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.472954 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="sg-core" containerID="cri-o://e6bde295c7812497b3848b236f4e958ec3015f8ebeca1a4fdbd6ee1413805cb9" gracePeriod=30 Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.473016 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="proxy-httpd" containerID="cri-o://2d6e21accb5b365373d4e446af9a0d2f3bb681b5f74d1a1d1d5bbeb21b98d345" gracePeriod=30 Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.573727 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.210:3000/\": read tcp 10.217.0.2:42148->10.217.0.210:3000: read: connection reset by peer" Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.594180 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.596529 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 12 18:25:50 crc kubenswrapper[4926]: I0312 18:25:50.601536 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.004125 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.317278 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" event={"ID":"99d97cdd-8bee-43b1-a07c-fee61fceff3a","Type":"ContainerStarted","Data":"c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c"} Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.317643 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.320869 4926 generic.go:334] "Generic (PLEG): container finished" podID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerID="2d6e21accb5b365373d4e446af9a0d2f3bb681b5f74d1a1d1d5bbeb21b98d345" exitCode=0 Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.320893 4926 generic.go:334] "Generic (PLEG): container finished" podID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerID="e6bde295c7812497b3848b236f4e958ec3015f8ebeca1a4fdbd6ee1413805cb9" exitCode=2 Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.320901 4926 generic.go:334] "Generic (PLEG): container finished" podID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerID="22a2d1e8587226f797d924ae278b1c0e7b846e0cc05d9f335cbb31ac04af937c" exitCode=0 Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.320908 4926 generic.go:334] "Generic (PLEG): container finished" podID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerID="00d08df89cba0281fd578a80d34eeb606178fe179a18e0bb4dca41d34fcaae6c" exitCode=0 Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.321072 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-log" containerID="cri-o://a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12" gracePeriod=30 Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.321282 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerDied","Data":"2d6e21accb5b365373d4e446af9a0d2f3bb681b5f74d1a1d1d5bbeb21b98d345"} Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.321330 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerDied","Data":"e6bde295c7812497b3848b236f4e958ec3015f8ebeca1a4fdbd6ee1413805cb9"} Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.321340 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerDied","Data":"22a2d1e8587226f797d924ae278b1c0e7b846e0cc05d9f335cbb31ac04af937c"} Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.321349 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerDied","Data":"00d08df89cba0281fd578a80d34eeb606178fe179a18e0bb4dca41d34fcaae6c"} Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.321357 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd8b5fe-8294-40c2-a6f4-3b82012686ec","Type":"ContainerDied","Data":"da79f6ead62afe8afb991cfd28d34557304f0f537134c344d4948f2cb4bfd386"} Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.321366 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da79f6ead62afe8afb991cfd28d34557304f0f537134c344d4948f2cb4bfd386" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.322594 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-api" containerID="cri-o://6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c" gracePeriod=30 Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.332202 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.342014 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" podStartSLOduration=3.341997774 podStartE2EDuration="3.341997774s" podCreationTimestamp="2026-03-12 18:25:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:51.337871535 +0000 UTC m=+1391.706497868" watchObservedRunningTime="2026-03-12 18:25:51.341997774 +0000 UTC m=+1391.710624107" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.353969 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383135 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-ceilometer-tls-certs\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383218 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-combined-ca-bundle\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383283 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-run-httpd\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383305 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-sg-core-conf-yaml\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383323 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-log-httpd\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383347 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-scripts\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383370 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh6lh\" (UniqueName: \"kubernetes.io/projected/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-kube-api-access-hh6lh\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.383474 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-config-data\") pod \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\" (UID: \"ccd8b5fe-8294-40c2-a6f4-3b82012686ec\") " Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.395596 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.395675 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.398428 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-scripts" (OuterVolumeSpecName: "scripts") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.402883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-kube-api-access-hh6lh" (OuterVolumeSpecName: "kube-api-access-hh6lh") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "kube-api-access-hh6lh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.442917 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.461407 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.485455 4926 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.485686 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.485775 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.485863 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.485944 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-scripts\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.486024 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh6lh\" (UniqueName: \"kubernetes.io/projected/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-kube-api-access-hh6lh\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.497418 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.541195 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-config-data" (OuterVolumeSpecName: "config-data") pod "ccd8b5fe-8294-40c2-a6f4-3b82012686ec" (UID: "ccd8b5fe-8294-40c2-a6f4-3b82012686ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.587179 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:51 crc kubenswrapper[4926]: I0312 18:25:51.587209 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd8b5fe-8294-40c2-a6f4-3b82012686ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.329698 4926 generic.go:334] "Generic (PLEG): container finished" podID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerID="a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12" exitCode=143 Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.329793 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4","Type":"ContainerDied","Data":"a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12"} Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.330042 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.361255 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.371243 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.392336 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:52 crc kubenswrapper[4926]: E0312 18:25:52.392819 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-notification-agent" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.392843 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-notification-agent" Mar 12 18:25:52 crc kubenswrapper[4926]: E0312 18:25:52.392863 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-central-agent" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.392872 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-central-agent" Mar 12 18:25:52 crc kubenswrapper[4926]: E0312 18:25:52.392896 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="proxy-httpd" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.392905 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="proxy-httpd" Mar 12 18:25:52 crc kubenswrapper[4926]: E0312 18:25:52.392935 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="sg-core" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.392944 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="sg-core" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.393163 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-central-agent" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.393177 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="sg-core" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.393193 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="proxy-httpd" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.393203 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" containerName="ceilometer-notification-agent" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.394997 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.397009 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.397536 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.397732 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.416839 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.500199 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd8b5fe-8294-40c2-a6f4-3b82012686ec" path="/var/lib/kubelet/pods/ccd8b5fe-8294-40c2-a6f4-3b82012686ec/volumes" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.504905 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.504969 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.505029 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.505169 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-scripts\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.505277 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f983b88e-aba3-4d49-bbd4-4db5eef5266c-run-httpd\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.505320 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f983b88e-aba3-4d49-bbd4-4db5eef5266c-log-httpd\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.505341 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-config-data\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.505398 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6fb7\" (UniqueName: \"kubernetes.io/projected/f983b88e-aba3-4d49-bbd4-4db5eef5266c-kube-api-access-h6fb7\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607099 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-scripts\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607186 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f983b88e-aba3-4d49-bbd4-4db5eef5266c-run-httpd\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607221 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f983b88e-aba3-4d49-bbd4-4db5eef5266c-log-httpd\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607236 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-config-data\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607267 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6fb7\" (UniqueName: \"kubernetes.io/projected/f983b88e-aba3-4d49-bbd4-4db5eef5266c-kube-api-access-h6fb7\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607308 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607358 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.607378 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.608629 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f983b88e-aba3-4d49-bbd4-4db5eef5266c-log-httpd\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.608941 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f983b88e-aba3-4d49-bbd4-4db5eef5266c-run-httpd\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.613549 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.613658 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.614159 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-config-data\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.626024 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-scripts\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.630228 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6fb7\" (UniqueName: \"kubernetes.io/projected/f983b88e-aba3-4d49-bbd4-4db5eef5266c-kube-api-access-h6fb7\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.638148 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f983b88e-aba3-4d49-bbd4-4db5eef5266c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f983b88e-aba3-4d49-bbd4-4db5eef5266c\") " pod="openstack/ceilometer-0" Mar 12 18:25:52 crc kubenswrapper[4926]: I0312 18:25:52.743827 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 12 18:25:53 crc kubenswrapper[4926]: I0312 18:25:53.268039 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 12 18:25:53 crc kubenswrapper[4926]: I0312 18:25:53.344106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f983b88e-aba3-4d49-bbd4-4db5eef5266c","Type":"ContainerStarted","Data":"efe78a1ebaf76bd5ca4c5c7d2389c61f86fefc1d3ee135c5723dc145b1fb67ee"} Mar 12 18:25:54 crc kubenswrapper[4926]: I0312 18:25:54.940740 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.084874 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-logs\") pod \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.084945 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-combined-ca-bundle\") pod \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.084992 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsrb4\" (UniqueName: \"kubernetes.io/projected/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-kube-api-access-qsrb4\") pod \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.085195 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-config-data\") pod \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\" (UID: \"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4\") " Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.085642 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-logs" (OuterVolumeSpecName: "logs") pod "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" (UID: "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.085857 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-logs\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.096418 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-kube-api-access-qsrb4" (OuterVolumeSpecName: "kube-api-access-qsrb4") pod "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" (UID: "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4"). InnerVolumeSpecName "kube-api-access-qsrb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.121627 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-config-data" (OuterVolumeSpecName: "config-data") pod "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" (UID: "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.121735 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" (UID: "d37e2fa1-9044-4dfb-9aea-bcb441cc91a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.187904 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.187936 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.187947 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsrb4\" (UniqueName: \"kubernetes.io/projected/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4-kube-api-access-qsrb4\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.363016 4926 generic.go:334] "Generic (PLEG): container finished" podID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerID="6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c" exitCode=0 Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.363096 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.363106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4","Type":"ContainerDied","Data":"6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c"} Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.363584 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d37e2fa1-9044-4dfb-9aea-bcb441cc91a4","Type":"ContainerDied","Data":"e038299ff9160ecc8a030d680499ae9a217c62acd0df47033360941b4347d3a1"} Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.363606 4926 scope.go:117] "RemoveContainer" containerID="6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.390677 4926 scope.go:117] "RemoveContainer" containerID="a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.417905 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.418150 4926 scope.go:117] "RemoveContainer" containerID="6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c" Mar 12 18:25:55 crc kubenswrapper[4926]: E0312 18:25:55.422118 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c\": container with ID starting with 6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c not found: ID does not exist" containerID="6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.422163 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c"} err="failed to get container status \"6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c\": rpc error: code = NotFound desc = could not find container \"6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c\": container with ID starting with 6fa6d015869bdf6e7a8cb47337d72a2bc01b6eda4f68a9902273a561f45b686c not found: ID does not exist" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.422187 4926 scope.go:117] "RemoveContainer" containerID="a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12" Mar 12 18:25:55 crc kubenswrapper[4926]: E0312 18:25:55.422451 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12\": container with ID starting with a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12 not found: ID does not exist" containerID="a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.422494 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12"} err="failed to get container status \"a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12\": rpc error: code = NotFound desc = could not find container \"a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12\": container with ID starting with a1113ff82e09dd62b2b62d25d27760593aaf0e686a84b599e8ba5e2556f57a12 not found: ID does not exist" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.431964 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.463728 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:55 crc kubenswrapper[4926]: E0312 18:25:55.464284 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-api" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.464310 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-api" Mar 12 18:25:55 crc kubenswrapper[4926]: E0312 18:25:55.464342 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-log" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.464350 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-log" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.464574 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-log" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.464613 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" containerName="nova-api-api" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.465802 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.468101 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.468477 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.472749 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.479479 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.595352 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.595476 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-config-data\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.595502 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r22k\" (UniqueName: \"kubernetes.io/projected/65b09111-c033-45e3-97d3-cd755e1a79ab-kube-api-access-9r22k\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.595615 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.595714 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-public-tls-certs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.595747 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65b09111-c033-45e3-97d3-cd755e1a79ab-logs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.698623 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65b09111-c033-45e3-97d3-cd755e1a79ab-logs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.698743 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.698799 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-config-data\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.698820 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r22k\" (UniqueName: \"kubernetes.io/projected/65b09111-c033-45e3-97d3-cd755e1a79ab-kube-api-access-9r22k\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.698919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.698972 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-public-tls-certs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.699172 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65b09111-c033-45e3-97d3-cd755e1a79ab-logs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.702710 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.702807 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.703083 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-config-data\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.703485 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65b09111-c033-45e3-97d3-cd755e1a79ab-public-tls-certs\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.718093 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r22k\" (UniqueName: \"kubernetes.io/projected/65b09111-c033-45e3-97d3-cd755e1a79ab-kube-api-access-9r22k\") pod \"nova-api-0\" (UID: \"65b09111-c033-45e3-97d3-cd755e1a79ab\") " pod="openstack/nova-api-0" Mar 12 18:25:55 crc kubenswrapper[4926]: I0312 18:25:55.788027 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 12 18:25:56 crc kubenswrapper[4926]: I0312 18:25:56.253542 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 12 18:25:56 crc kubenswrapper[4926]: I0312 18:25:56.373280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65b09111-c033-45e3-97d3-cd755e1a79ab","Type":"ContainerStarted","Data":"ca502a96f6aafefcf167c3f572b8801d7bb44e17206f063d0973b12d38d5dc86"} Mar 12 18:25:56 crc kubenswrapper[4926]: I0312 18:25:56.503240 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d37e2fa1-9044-4dfb-9aea-bcb441cc91a4" path="/var/lib/kubelet/pods/d37e2fa1-9044-4dfb-9aea-bcb441cc91a4/volumes" Mar 12 18:25:57 crc kubenswrapper[4926]: I0312 18:25:57.393009 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65b09111-c033-45e3-97d3-cd755e1a79ab","Type":"ContainerStarted","Data":"3d2fbf4f1a9d4f7d8849384d3e26061ddbfabbfd65065d6a6fc041ee4c650106"} Mar 12 18:25:57 crc kubenswrapper[4926]: I0312 18:25:57.393866 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65b09111-c033-45e3-97d3-cd755e1a79ab","Type":"ContainerStarted","Data":"cbfe1612678ced632699b1895e35ffde3fc90098b673e8dcf8c6607ca19a1e86"} Mar 12 18:25:57 crc kubenswrapper[4926]: I0312 18:25:57.430686 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.4306603190000002 podStartE2EDuration="2.430660319s" podCreationTimestamp="2026-03-12 18:25:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:25:57.416393941 +0000 UTC m=+1397.785020304" watchObservedRunningTime="2026-03-12 18:25:57.430660319 +0000 UTC m=+1397.799286682" Mar 12 18:25:58 crc kubenswrapper[4926]: I0312 18:25:58.406077 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f983b88e-aba3-4d49-bbd4-4db5eef5266c","Type":"ContainerStarted","Data":"f7f6ae8fa45482b7b8fa1323c6a6628ac7710b7c51525caebc28a710eed57fd0"} Mar 12 18:25:58 crc kubenswrapper[4926]: I0312 18:25:58.860634 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:25:58 crc kubenswrapper[4926]: I0312 18:25:58.921600 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwx72"] Mar 12 18:25:58 crc kubenswrapper[4926]: I0312 18:25:58.921842 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" podUID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerName="dnsmasq-dns" containerID="cri-o://9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3" gracePeriod=10 Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.413041 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.422061 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f983b88e-aba3-4d49-bbd4-4db5eef5266c","Type":"ContainerStarted","Data":"30fe6bd671ca7ba78619b5143656473d3732a17f263f07921efaad240adbb7e2"} Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.422150 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f983b88e-aba3-4d49-bbd4-4db5eef5266c","Type":"ContainerStarted","Data":"8833b4e39e8c00c9608f9b821b7134299d5fb3f79586aecd64a3adf71f01a68e"} Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.425191 4926 generic.go:334] "Generic (PLEG): container finished" podID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerID="9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3" exitCode=0 Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.425255 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" event={"ID":"458c7cf7-4e8a-4272-8940-1d730293a0ca","Type":"ContainerDied","Data":"9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3"} Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.425310 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" event={"ID":"458c7cf7-4e8a-4272-8940-1d730293a0ca","Type":"ContainerDied","Data":"820539f8ae4b6d013f20317aef2ceed339b3e7e027d5738aa8b884143c185f6f"} Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.425333 4926 scope.go:117] "RemoveContainer" containerID="9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.425817 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-jwx72" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.457857 4926 scope.go:117] "RemoveContainer" containerID="12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.480992 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-sb\") pod \"458c7cf7-4e8a-4272-8940-1d730293a0ca\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.481040 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-nb\") pod \"458c7cf7-4e8a-4272-8940-1d730293a0ca\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.481098 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7xnp\" (UniqueName: \"kubernetes.io/projected/458c7cf7-4e8a-4272-8940-1d730293a0ca-kube-api-access-k7xnp\") pod \"458c7cf7-4e8a-4272-8940-1d730293a0ca\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.481173 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-svc\") pod \"458c7cf7-4e8a-4272-8940-1d730293a0ca\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.482303 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-config\") pod \"458c7cf7-4e8a-4272-8940-1d730293a0ca\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.482764 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-swift-storage-0\") pod \"458c7cf7-4e8a-4272-8940-1d730293a0ca\" (UID: \"458c7cf7-4e8a-4272-8940-1d730293a0ca\") " Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.484887 4926 scope.go:117] "RemoveContainer" containerID="9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3" Mar 12 18:25:59 crc kubenswrapper[4926]: E0312 18:25:59.485295 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3\": container with ID starting with 9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3 not found: ID does not exist" containerID="9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.485323 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3"} err="failed to get container status \"9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3\": rpc error: code = NotFound desc = could not find container \"9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3\": container with ID starting with 9aea8ffd7746f95a03c2a7236be4f05b709c69b3e443e5e730cc0dfa7993bfc3 not found: ID does not exist" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.485344 4926 scope.go:117] "RemoveContainer" containerID="12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491" Mar 12 18:25:59 crc kubenswrapper[4926]: E0312 18:25:59.485852 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491\": container with ID starting with 12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491 not found: ID does not exist" containerID="12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.485871 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491"} err="failed to get container status \"12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491\": rpc error: code = NotFound desc = could not find container \"12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491\": container with ID starting with 12cbd32c692806fa4d603607cb6ae7ddf7f80a9712bcefaec0094fde0defb491 not found: ID does not exist" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.489919 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458c7cf7-4e8a-4272-8940-1d730293a0ca-kube-api-access-k7xnp" (OuterVolumeSpecName: "kube-api-access-k7xnp") pod "458c7cf7-4e8a-4272-8940-1d730293a0ca" (UID: "458c7cf7-4e8a-4272-8940-1d730293a0ca"). InnerVolumeSpecName "kube-api-access-k7xnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.536228 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "458c7cf7-4e8a-4272-8940-1d730293a0ca" (UID: "458c7cf7-4e8a-4272-8940-1d730293a0ca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.543954 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "458c7cf7-4e8a-4272-8940-1d730293a0ca" (UID: "458c7cf7-4e8a-4272-8940-1d730293a0ca"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.544190 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "458c7cf7-4e8a-4272-8940-1d730293a0ca" (UID: "458c7cf7-4e8a-4272-8940-1d730293a0ca"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.545598 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-config" (OuterVolumeSpecName: "config") pod "458c7cf7-4e8a-4272-8940-1d730293a0ca" (UID: "458c7cf7-4e8a-4272-8940-1d730293a0ca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.550537 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "458c7cf7-4e8a-4272-8940-1d730293a0ca" (UID: "458c7cf7-4e8a-4272-8940-1d730293a0ca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.584580 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7xnp\" (UniqueName: \"kubernetes.io/projected/458c7cf7-4e8a-4272-8940-1d730293a0ca-kube-api-access-k7xnp\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.584619 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.584629 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.584640 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.584649 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.584658 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/458c7cf7-4e8a-4272-8940-1d730293a0ca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.762077 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwx72"] Mar 12 18:25:59 crc kubenswrapper[4926]: I0312 18:25:59.769349 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-jwx72"] Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.137708 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555666-x5zgg"] Mar 12 18:26:00 crc kubenswrapper[4926]: E0312 18:26:00.138413 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerName="init" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.138429 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerName="init" Mar 12 18:26:00 crc kubenswrapper[4926]: E0312 18:26:00.138477 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerName="dnsmasq-dns" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.138485 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerName="dnsmasq-dns" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.138690 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="458c7cf7-4e8a-4272-8940-1d730293a0ca" containerName="dnsmasq-dns" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.139407 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555666-x5zgg" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.141951 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.142265 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.143038 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.147579 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555666-x5zgg"] Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.195636 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj8b2\" (UniqueName: \"kubernetes.io/projected/5ff186e2-5cbe-493a-b911-426e982888cb-kube-api-access-fj8b2\") pod \"auto-csr-approver-29555666-x5zgg\" (UID: \"5ff186e2-5cbe-493a-b911-426e982888cb\") " pod="openshift-infra/auto-csr-approver-29555666-x5zgg" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.298122 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj8b2\" (UniqueName: \"kubernetes.io/projected/5ff186e2-5cbe-493a-b911-426e982888cb-kube-api-access-fj8b2\") pod \"auto-csr-approver-29555666-x5zgg\" (UID: \"5ff186e2-5cbe-493a-b911-426e982888cb\") " pod="openshift-infra/auto-csr-approver-29555666-x5zgg" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.322523 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj8b2\" (UniqueName: \"kubernetes.io/projected/5ff186e2-5cbe-493a-b911-426e982888cb-kube-api-access-fj8b2\") pod \"auto-csr-approver-29555666-x5zgg\" (UID: \"5ff186e2-5cbe-493a-b911-426e982888cb\") " pod="openshift-infra/auto-csr-approver-29555666-x5zgg" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.470657 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555666-x5zgg" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.557349 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="458c7cf7-4e8a-4272-8940-1d730293a0ca" path="/var/lib/kubelet/pods/458c7cf7-4e8a-4272-8940-1d730293a0ca/volumes" Mar 12 18:26:00 crc kubenswrapper[4926]: I0312 18:26:00.994173 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555666-x5zgg"] Mar 12 18:26:01 crc kubenswrapper[4926]: W0312 18:26:01.008894 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ff186e2_5cbe_493a_b911_426e982888cb.slice/crio-c1f487ad4c77fc31fb454e4d9a71da5681d79d268758afebdd45515eec4de874 WatchSource:0}: Error finding container c1f487ad4c77fc31fb454e4d9a71da5681d79d268758afebdd45515eec4de874: Status 404 returned error can't find the container with id c1f487ad4c77fc31fb454e4d9a71da5681d79d268758afebdd45515eec4de874 Mar 12 18:26:01 crc kubenswrapper[4926]: I0312 18:26:01.450764 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f983b88e-aba3-4d49-bbd4-4db5eef5266c","Type":"ContainerStarted","Data":"bac49af58acca4e456a30754fa1a58fd442847e42e37e4e2d91dcb1b439fd8e0"} Mar 12 18:26:01 crc kubenswrapper[4926]: I0312 18:26:01.451123 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 12 18:26:01 crc kubenswrapper[4926]: I0312 18:26:01.452477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555666-x5zgg" event={"ID":"5ff186e2-5cbe-493a-b911-426e982888cb","Type":"ContainerStarted","Data":"c1f487ad4c77fc31fb454e4d9a71da5681d79d268758afebdd45515eec4de874"} Mar 12 18:26:01 crc kubenswrapper[4926]: I0312 18:26:01.501247 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.589875358 podStartE2EDuration="9.501213248s" podCreationTimestamp="2026-03-12 18:25:52 +0000 UTC" firstStartedPulling="2026-03-12 18:25:53.270075114 +0000 UTC m=+1393.638701457" lastFinishedPulling="2026-03-12 18:26:01.181413014 +0000 UTC m=+1401.550039347" observedRunningTime="2026-03-12 18:26:01.485674211 +0000 UTC m=+1401.854300544" watchObservedRunningTime="2026-03-12 18:26:01.501213248 +0000 UTC m=+1401.869839601" Mar 12 18:26:03 crc kubenswrapper[4926]: I0312 18:26:03.476002 4926 generic.go:334] "Generic (PLEG): container finished" podID="5ff186e2-5cbe-493a-b911-426e982888cb" containerID="303e0f37b5ff32adc0c5ce3a0b0f3252de8de58704f569f8ef9bdf6d235b21d1" exitCode=0 Mar 12 18:26:03 crc kubenswrapper[4926]: I0312 18:26:03.476077 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555666-x5zgg" event={"ID":"5ff186e2-5cbe-493a-b911-426e982888cb","Type":"ContainerDied","Data":"303e0f37b5ff32adc0c5ce3a0b0f3252de8de58704f569f8ef9bdf6d235b21d1"} Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:04.907590 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555666-x5zgg" Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.022618 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj8b2\" (UniqueName: \"kubernetes.io/projected/5ff186e2-5cbe-493a-b911-426e982888cb-kube-api-access-fj8b2\") pod \"5ff186e2-5cbe-493a-b911-426e982888cb\" (UID: \"5ff186e2-5cbe-493a-b911-426e982888cb\") " Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.037826 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ff186e2-5cbe-493a-b911-426e982888cb-kube-api-access-fj8b2" (OuterVolumeSpecName: "kube-api-access-fj8b2") pod "5ff186e2-5cbe-493a-b911-426e982888cb" (UID: "5ff186e2-5cbe-493a-b911-426e982888cb"). InnerVolumeSpecName "kube-api-access-fj8b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.125099 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj8b2\" (UniqueName: \"kubernetes.io/projected/5ff186e2-5cbe-493a-b911-426e982888cb-kube-api-access-fj8b2\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.502716 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555666-x5zgg" event={"ID":"5ff186e2-5cbe-493a-b911-426e982888cb","Type":"ContainerDied","Data":"c1f487ad4c77fc31fb454e4d9a71da5681d79d268758afebdd45515eec4de874"} Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.503103 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1f487ad4c77fc31fb454e4d9a71da5681d79d268758afebdd45515eec4de874" Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.503170 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555666-x5zgg" Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.789027 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.789111 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.980194 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555660-cbx5b"] Mar 12 18:26:05 crc kubenswrapper[4926]: I0312 18:26:05.992198 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555660-cbx5b"] Mar 12 18:26:06 crc kubenswrapper[4926]: I0312 18:26:06.509275 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b50fb579-57d6-4029-a4f3-c8a3303bac4d" path="/var/lib/kubelet/pods/b50fb579-57d6-4029-a4f3-c8a3303bac4d/volumes" Mar 12 18:26:06 crc kubenswrapper[4926]: I0312 18:26:06.800580 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65b09111-c033-45e3-97d3-cd755e1a79ab" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.215:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 18:26:06 crc kubenswrapper[4926]: I0312 18:26:06.800650 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65b09111-c033-45e3-97d3-cd755e1a79ab" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.215:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 12 18:26:14 crc kubenswrapper[4926]: I0312 18:26:14.202205 4926 scope.go:117] "RemoveContainer" containerID="6f2e2b241a59eb4ec9fee306fab9cd670167a7fcd068ac1d7f97bc699a2c0ee6" Mar 12 18:26:15 crc kubenswrapper[4926]: I0312 18:26:15.797335 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 18:26:15 crc kubenswrapper[4926]: I0312 18:26:15.798023 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 18:26:15 crc kubenswrapper[4926]: I0312 18:26:15.802073 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 12 18:26:15 crc kubenswrapper[4926]: I0312 18:26:15.813256 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 18:26:16 crc kubenswrapper[4926]: I0312 18:26:16.614331 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 12 18:26:16 crc kubenswrapper[4926]: I0312 18:26:16.626928 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 12 18:26:22 crc kubenswrapper[4926]: I0312 18:26:22.762290 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 12 18:26:31 crc kubenswrapper[4926]: I0312 18:26:31.677341 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:26:33 crc kubenswrapper[4926]: I0312 18:26:33.509528 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:26:35 crc kubenswrapper[4926]: I0312 18:26:35.779090 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerName="rabbitmq" containerID="cri-o://ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a" gracePeriod=604796 Mar 12 18:26:37 crc kubenswrapper[4926]: I0312 18:26:37.483270 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerName="rabbitmq" containerID="cri-o://3e4a5dd026300b0565c4a11203b217973524caa9bcc9839e71d49583b866e88e" gracePeriod=604797 Mar 12 18:26:37 crc kubenswrapper[4926]: I0312 18:26:37.504078 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Mar 12 18:26:37 crc kubenswrapper[4926]: I0312 18:26:37.770388 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.430626 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.635915 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-plugins-conf\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.635969 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-config-data\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636021 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7kq2\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-kube-api-access-x7kq2\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636162 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-plugins\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636239 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c04aaec-485d-492f-8c24-e6860d9c78f7-pod-info\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636303 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-confd\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636359 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636391 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-erlang-cookie\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636470 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-server-conf\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636499 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c04aaec-485d-492f-8c24-e6860d9c78f7-erlang-cookie-secret\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.636539 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-tls\") pod \"9c04aaec-485d-492f-8c24-e6860d9c78f7\" (UID: \"9c04aaec-485d-492f-8c24-e6860d9c78f7\") " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.639721 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.645497 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.646985 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9c04aaec-485d-492f-8c24-e6860d9c78f7-pod-info" (OuterVolumeSpecName: "pod-info") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.647019 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-kube-api-access-x7kq2" (OuterVolumeSpecName: "kube-api-access-x7kq2") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "kube-api-access-x7kq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.647759 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.652516 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.656402 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c04aaec-485d-492f-8c24-e6860d9c78f7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.689563 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.692392 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-config-data" (OuterVolumeSpecName: "config-data") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.723349 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-server-conf" (OuterVolumeSpecName: "server-conf") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740195 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740225 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740237 4926 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-server-conf\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740247 4926 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9c04aaec-485d-492f-8c24-e6860d9c78f7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740288 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740297 4926 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740308 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9c04aaec-485d-492f-8c24-e6860d9c78f7-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740317 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7kq2\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-kube-api-access-x7kq2\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740326 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.740334 4926 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9c04aaec-485d-492f-8c24-e6860d9c78f7-pod-info\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.766080 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.794661 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9c04aaec-485d-492f-8c24-e6860d9c78f7" (UID: "9c04aaec-485d-492f-8c24-e6860d9c78f7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.840933 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9c04aaec-485d-492f-8c24-e6860d9c78f7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.840969 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.920698 4926 generic.go:334] "Generic (PLEG): container finished" podID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerID="ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a" exitCode=0 Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.920777 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.920788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9c04aaec-485d-492f-8c24-e6860d9c78f7","Type":"ContainerDied","Data":"ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a"} Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.920837 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9c04aaec-485d-492f-8c24-e6860d9c78f7","Type":"ContainerDied","Data":"1dfe0defccd325a5f2aba559c2e683f83b8337f73cdd57e5ea2b609807eb38bf"} Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.920855 4926 scope.go:117] "RemoveContainer" containerID="ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.944315 4926 scope.go:117] "RemoveContainer" containerID="68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.963883 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.971248 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.984100 4926 scope.go:117] "RemoveContainer" containerID="ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a" Mar 12 18:26:42 crc kubenswrapper[4926]: E0312 18:26:42.984707 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a\": container with ID starting with ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a not found: ID does not exist" containerID="ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.984737 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a"} err="failed to get container status \"ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a\": rpc error: code = NotFound desc = could not find container \"ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a\": container with ID starting with ece7b3f0f166f5f8791a6a655bf54bcbe43cea3c1a722f7e373529011b357a1a not found: ID does not exist" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.984762 4926 scope.go:117] "RemoveContainer" containerID="68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5" Mar 12 18:26:42 crc kubenswrapper[4926]: E0312 18:26:42.985179 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5\": container with ID starting with 68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5 not found: ID does not exist" containerID="68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.985209 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5"} err="failed to get container status \"68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5\": rpc error: code = NotFound desc = could not find container \"68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5\": container with ID starting with 68d563b06d81d99aae7648289259410aed203cf265688cfbf80718ef20c0aca5 not found: ID does not exist" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.999348 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:26:42 crc kubenswrapper[4926]: E0312 18:26:42.999844 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff186e2-5cbe-493a-b911-426e982888cb" containerName="oc" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.999867 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff186e2-5cbe-493a-b911-426e982888cb" containerName="oc" Mar 12 18:26:42 crc kubenswrapper[4926]: E0312 18:26:42.999894 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerName="setup-container" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.999905 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerName="setup-container" Mar 12 18:26:42 crc kubenswrapper[4926]: E0312 18:26:42.999948 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerName="rabbitmq" Mar 12 18:26:42 crc kubenswrapper[4926]: I0312 18:26:42.999957 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerName="rabbitmq" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.000195 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" containerName="rabbitmq" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.000216 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff186e2-5cbe-493a-b911-426e982888cb" containerName="oc" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.001733 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.005176 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.005476 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.005618 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-k928p" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.005659 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.005689 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.005622 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.005629 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.021418 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.149017 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.149106 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.149150 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.149245 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9aba0434-585e-4355-8019-1612400b2350-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.149590 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7px9q\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-kube-api-access-7px9q\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.149711 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-config-data\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.149831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.150124 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.150376 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9aba0434-585e-4355-8019-1612400b2350-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.150462 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.150536 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252617 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252689 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9aba0434-585e-4355-8019-1612400b2350-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252754 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252772 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252793 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252809 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252834 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252853 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9aba0434-585e-4355-8019-1612400b2350-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252894 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7px9q\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-kube-api-access-7px9q\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.252913 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-config-data\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.253891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-config-data\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.254964 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-server-conf\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.255857 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9aba0434-585e-4355-8019-1612400b2350-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.256499 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.256902 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.257623 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.260361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.259651 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9aba0434-585e-4355-8019-1612400b2350-pod-info\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.264283 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9aba0434-585e-4355-8019-1612400b2350-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.266250 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.281151 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7px9q\" (UniqueName: \"kubernetes.io/projected/9aba0434-585e-4355-8019-1612400b2350-kube-api-access-7px9q\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.301889 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"9aba0434-585e-4355-8019-1612400b2350\") " pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.387012 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.884502 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.932485 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9aba0434-585e-4355-8019-1612400b2350","Type":"ContainerStarted","Data":"bdc479238c3510dba7a47c2f6b2e81b6d4896fc6339ded89c262931f2379255c"} Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.934512 4926 generic.go:334] "Generic (PLEG): container finished" podID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerID="3e4a5dd026300b0565c4a11203b217973524caa9bcc9839e71d49583b866e88e" exitCode=0 Mar 12 18:26:43 crc kubenswrapper[4926]: I0312 18:26:43.934602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06f09c04-6c8d-4c47-a0a5-59def6ebbf94","Type":"ContainerDied","Data":"3e4a5dd026300b0565c4a11203b217973524caa9bcc9839e71d49583b866e88e"} Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.212700 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379258 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-pod-info\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379646 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379679 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-erlang-cookie-secret\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379727 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-server-conf\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379765 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-erlang-cookie\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379803 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kzhl\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-kube-api-access-8kzhl\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379858 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-tls\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379942 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-plugins-conf\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.379972 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-confd\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.380007 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-plugins\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.380057 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-config-data\") pod \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\" (UID: \"06f09c04-6c8d-4c47-a0a5-59def6ebbf94\") " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.385079 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.385355 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.388265 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.388682 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-kube-api-access-8kzhl" (OuterVolumeSpecName: "kube-api-access-8kzhl") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "kube-api-access-8kzhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.389628 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.392573 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-pod-info" (OuterVolumeSpecName: "pod-info") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.392705 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.403465 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.439303 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-config-data" (OuterVolumeSpecName: "config-data") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.466777 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-server-conf" (OuterVolumeSpecName: "server-conf") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481868 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481898 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-config-data\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481908 4926 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-pod-info\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481935 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481946 4926 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481957 4926 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-server-conf\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481966 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481977 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kzhl\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-kube-api-access-8kzhl\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481985 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.481993 4926 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.502190 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.502256 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c04aaec-485d-492f-8c24-e6860d9c78f7" path="/var/lib/kubelet/pods/9c04aaec-485d-492f-8c24-e6860d9c78f7/volumes" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.546881 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "06f09c04-6c8d-4c47-a0a5-59def6ebbf94" (UID: "06f09c04-6c8d-4c47-a0a5-59def6ebbf94"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.584052 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06f09c04-6c8d-4c47-a0a5-59def6ebbf94-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.584086 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.955661 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06f09c04-6c8d-4c47-a0a5-59def6ebbf94","Type":"ContainerDied","Data":"851881376da52e23e91e40413019b432ba8d1a58a75cb35642884721899d9a50"} Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.955723 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:44 crc kubenswrapper[4926]: I0312 18:26:44.955735 4926 scope.go:117] "RemoveContainer" containerID="3e4a5dd026300b0565c4a11203b217973524caa9bcc9839e71d49583b866e88e" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.005647 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.026502 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.075203 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:26:45 crc kubenswrapper[4926]: E0312 18:26:45.075670 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerName="rabbitmq" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.075685 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerName="rabbitmq" Mar 12 18:26:45 crc kubenswrapper[4926]: E0312 18:26:45.075699 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerName="setup-container" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.075705 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerName="setup-container" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.075858 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" containerName="rabbitmq" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.087922 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.090585 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.090797 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.094029 4926 scope.go:117] "RemoveContainer" containerID="cfa97fbf75b6c2e852f5d66a227450ea6c8da1b0641deb3da6fb3c35ba9a0f12" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.097344 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.097526 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.102867 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.103237 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-r4knd" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.103480 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.103667 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.197125 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9b82f03-7ac1-4805-858b-708760b4e476-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.197731 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.197902 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198006 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198104 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9b82f03-7ac1-4805-858b-708760b4e476-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198236 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198353 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198470 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198636 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rzmd\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-kube-api-access-6rzmd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.198717 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301029 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301290 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301498 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301612 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rzmd\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-kube-api-access-6rzmd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301744 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301860 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9b82f03-7ac1-4805-858b-708760b4e476-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.301984 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.302560 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.302826 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.303417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.303489 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.303078 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.303388 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.303496 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9b82f03-7ac1-4805-858b-708760b4e476-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.303742 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.304076 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9b82f03-7ac1-4805-858b-708760b4e476-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.470231 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.471848 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9b82f03-7ac1-4805-858b-708760b4e476-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.471855 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9b82f03-7ac1-4805-858b-708760b4e476-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.472116 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rzmd\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-kube-api-access-6rzmd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.474111 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9b82f03-7ac1-4805-858b-708760b4e476-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.538807 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b9b82f03-7ac1-4805-858b-708760b4e476\") " pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.716327 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:26:45 crc kubenswrapper[4926]: I0312 18:26:45.985106 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9aba0434-585e-4355-8019-1612400b2350","Type":"ContainerStarted","Data":"f428e986bfed0b99e85dec3bb535e7ec0c4eb9551d8016425ac735d81340c54d"} Mar 12 18:26:46 crc kubenswrapper[4926]: I0312 18:26:46.094778 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 12 18:26:46 crc kubenswrapper[4926]: I0312 18:26:46.504221 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f09c04-6c8d-4c47-a0a5-59def6ebbf94" path="/var/lib/kubelet/pods/06f09c04-6c8d-4c47-a0a5-59def6ebbf94/volumes" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.004881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b9b82f03-7ac1-4805-858b-708760b4e476","Type":"ContainerStarted","Data":"c71d575325dd99be617816314834cd889a5cd67eb4b3f3e33266b5be5f6a25d4"} Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.793067 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4ldnw"] Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.796608 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.799941 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.806159 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4ldnw"] Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.874696 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.874761 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.874801 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-config\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.874868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fvxr\" (UniqueName: \"kubernetes.io/projected/ce48eb28-755b-4a18-8bf1-212baff1f7fc-kube-api-access-5fvxr\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.874946 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.874978 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.875044 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.977429 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.977860 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.977971 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.978064 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-config\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.978173 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fvxr\" (UniqueName: \"kubernetes.io/projected/ce48eb28-755b-4a18-8bf1-212baff1f7fc-kube-api-access-5fvxr\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.978296 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.978387 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.980033 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.980891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.982790 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-config\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.983293 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.984149 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:47 crc kubenswrapper[4926]: I0312 18:26:47.986070 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:48 crc kubenswrapper[4926]: I0312 18:26:48.037705 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fvxr\" (UniqueName: \"kubernetes.io/projected/ce48eb28-755b-4a18-8bf1-212baff1f7fc-kube-api-access-5fvxr\") pod \"dnsmasq-dns-79bd4cc8c9-4ldnw\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:48 crc kubenswrapper[4926]: I0312 18:26:48.045999 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b9b82f03-7ac1-4805-858b-708760b4e476","Type":"ContainerStarted","Data":"b958f13a715a00e9a33abc7ba17caf342fd31d1ad60a10155cff9bad1045c238"} Mar 12 18:26:48 crc kubenswrapper[4926]: I0312 18:26:48.116644 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:48 crc kubenswrapper[4926]: I0312 18:26:48.618866 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4ldnw"] Mar 12 18:26:49 crc kubenswrapper[4926]: I0312 18:26:49.071593 4926 generic.go:334] "Generic (PLEG): container finished" podID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerID="5c57deb5d6d68579f57781cc73b1f95fd91c467933afbbc36484489589a3e08d" exitCode=0 Mar 12 18:26:49 crc kubenswrapper[4926]: I0312 18:26:49.073133 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" event={"ID":"ce48eb28-755b-4a18-8bf1-212baff1f7fc","Type":"ContainerDied","Data":"5c57deb5d6d68579f57781cc73b1f95fd91c467933afbbc36484489589a3e08d"} Mar 12 18:26:49 crc kubenswrapper[4926]: I0312 18:26:49.073212 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" event={"ID":"ce48eb28-755b-4a18-8bf1-212baff1f7fc","Type":"ContainerStarted","Data":"73aeb25dea0852f705f0adfdaeaab49793c64b81144d0bee56d28bfe72283a50"} Mar 12 18:26:50 crc kubenswrapper[4926]: I0312 18:26:50.092789 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" event={"ID":"ce48eb28-755b-4a18-8bf1-212baff1f7fc","Type":"ContainerStarted","Data":"31005cd2c4f0038c1a2e2165529b7d457634779c21c8fccaed21553ed78e6d14"} Mar 12 18:26:50 crc kubenswrapper[4926]: I0312 18:26:50.093777 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:50 crc kubenswrapper[4926]: I0312 18:26:50.124195 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" podStartSLOduration=3.124165453 podStartE2EDuration="3.124165453s" podCreationTimestamp="2026-03-12 18:26:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:26:50.117928068 +0000 UTC m=+1450.486554431" watchObservedRunningTime="2026-03-12 18:26:50.124165453 +0000 UTC m=+1450.492791806" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.117766 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.190160 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-nfj99"] Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.190905 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" podUID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerName="dnsmasq-dns" containerID="cri-o://c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c" gracePeriod=10 Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.426149 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5fc8b56cf-6cbwx"] Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.427998 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.438653 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fc8b56cf-6cbwx"] Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.496164 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-dns-swift-storage-0\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.496226 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-openstack-edpm-ipam\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.496270 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-dns-svc\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.496309 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-ovsdbserver-nb\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.496803 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-config\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.497031 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-ovsdbserver-sb\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.497103 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trcwn\" (UniqueName: \"kubernetes.io/projected/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-kube-api-access-trcwn\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.602836 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-dns-swift-storage-0\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.603183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-openstack-edpm-ipam\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.603242 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-dns-svc\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.603284 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-ovsdbserver-nb\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.603349 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-config\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.603420 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-ovsdbserver-sb\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.603468 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trcwn\" (UniqueName: \"kubernetes.io/projected/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-kube-api-access-trcwn\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.604470 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-ovsdbserver-nb\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.605510 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-dns-swift-storage-0\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.606142 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-openstack-edpm-ipam\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.606201 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-config\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.606796 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-dns-svc\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.607358 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-ovsdbserver-sb\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.636234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trcwn\" (UniqueName: \"kubernetes.io/projected/6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69-kube-api-access-trcwn\") pod \"dnsmasq-dns-5fc8b56cf-6cbwx\" (UID: \"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69\") " pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.693263 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.704499 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-swift-storage-0\") pod \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.704592 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-sb\") pod \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.704672 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-config\") pod \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.704695 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-nb\") pod \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.704724 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-svc\") pod \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.704750 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swntz\" (UniqueName: \"kubernetes.io/projected/99d97cdd-8bee-43b1-a07c-fee61fceff3a-kube-api-access-swntz\") pod \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\" (UID: \"99d97cdd-8bee-43b1-a07c-fee61fceff3a\") " Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.709278 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d97cdd-8bee-43b1-a07c-fee61fceff3a-kube-api-access-swntz" (OuterVolumeSpecName: "kube-api-access-swntz") pod "99d97cdd-8bee-43b1-a07c-fee61fceff3a" (UID: "99d97cdd-8bee-43b1-a07c-fee61fceff3a"). InnerVolumeSpecName "kube-api-access-swntz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.758572 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.773230 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "99d97cdd-8bee-43b1-a07c-fee61fceff3a" (UID: "99d97cdd-8bee-43b1-a07c-fee61fceff3a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.795309 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99d97cdd-8bee-43b1-a07c-fee61fceff3a" (UID: "99d97cdd-8bee-43b1-a07c-fee61fceff3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.800136 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-config" (OuterVolumeSpecName: "config") pod "99d97cdd-8bee-43b1-a07c-fee61fceff3a" (UID: "99d97cdd-8bee-43b1-a07c-fee61fceff3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.807054 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.807086 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.807098 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.807106 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swntz\" (UniqueName: \"kubernetes.io/projected/99d97cdd-8bee-43b1-a07c-fee61fceff3a-kube-api-access-swntz\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.818668 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "99d97cdd-8bee-43b1-a07c-fee61fceff3a" (UID: "99d97cdd-8bee-43b1-a07c-fee61fceff3a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.826040 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "99d97cdd-8bee-43b1-a07c-fee61fceff3a" (UID: "99d97cdd-8bee-43b1-a07c-fee61fceff3a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.909562 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:58 crc kubenswrapper[4926]: I0312 18:26:58.909613 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99d97cdd-8bee-43b1-a07c-fee61fceff3a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.200230 4926 generic.go:334] "Generic (PLEG): container finished" podID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerID="c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c" exitCode=0 Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.200356 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.200390 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" event={"ID":"99d97cdd-8bee-43b1-a07c-fee61fceff3a","Type":"ContainerDied","Data":"c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c"} Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.200773 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-nfj99" event={"ID":"99d97cdd-8bee-43b1-a07c-fee61fceff3a","Type":"ContainerDied","Data":"9711fea2cdf454dba0540ad56623bf62d55c2c749c2e71e1fe9a2ccbb9165424"} Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.200810 4926 scope.go:117] "RemoveContainer" containerID="c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.214004 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5fc8b56cf-6cbwx"] Mar 12 18:26:59 crc kubenswrapper[4926]: W0312 18:26:59.218568 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f42a0ec_9b43_4a24_b5b2_89cd0f3abe69.slice/crio-7d27be680723fe4958ac9fa980d95153baa0ef070f295c6dfa217c24bf474a33 WatchSource:0}: Error finding container 7d27be680723fe4958ac9fa980d95153baa0ef070f295c6dfa217c24bf474a33: Status 404 returned error can't find the container with id 7d27be680723fe4958ac9fa980d95153baa0ef070f295c6dfa217c24bf474a33 Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.237174 4926 scope.go:117] "RemoveContainer" containerID="30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.258093 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-nfj99"] Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.264942 4926 scope.go:117] "RemoveContainer" containerID="c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.265003 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-nfj99"] Mar 12 18:26:59 crc kubenswrapper[4926]: E0312 18:26:59.265392 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c\": container with ID starting with c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c not found: ID does not exist" containerID="c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.265422 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c"} err="failed to get container status \"c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c\": rpc error: code = NotFound desc = could not find container \"c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c\": container with ID starting with c6990a64a13f798700cc9711674b7da102c1f60c04f410a014b3f9abaf48412c not found: ID does not exist" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.265516 4926 scope.go:117] "RemoveContainer" containerID="30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0" Mar 12 18:26:59 crc kubenswrapper[4926]: E0312 18:26:59.266016 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0\": container with ID starting with 30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0 not found: ID does not exist" containerID="30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0" Mar 12 18:26:59 crc kubenswrapper[4926]: I0312 18:26:59.266092 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0"} err="failed to get container status \"30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0\": rpc error: code = NotFound desc = could not find container \"30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0\": container with ID starting with 30acc030e2a3b017e634964990b6965f4e920c9d854391c2c49981117f81a7d0 not found: ID does not exist" Mar 12 18:27:00 crc kubenswrapper[4926]: I0312 18:27:00.211086 4926 generic.go:334] "Generic (PLEG): container finished" podID="6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69" containerID="e9a033282adcd73253cdaa2d902a610163965643025af868c4695b8f53713492" exitCode=0 Mar 12 18:27:00 crc kubenswrapper[4926]: I0312 18:27:00.211200 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" event={"ID":"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69","Type":"ContainerDied","Data":"e9a033282adcd73253cdaa2d902a610163965643025af868c4695b8f53713492"} Mar 12 18:27:00 crc kubenswrapper[4926]: I0312 18:27:00.211390 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" event={"ID":"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69","Type":"ContainerStarted","Data":"7d27be680723fe4958ac9fa980d95153baa0ef070f295c6dfa217c24bf474a33"} Mar 12 18:27:00 crc kubenswrapper[4926]: I0312 18:27:00.501447 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" path="/var/lib/kubelet/pods/99d97cdd-8bee-43b1-a07c-fee61fceff3a/volumes" Mar 12 18:27:01 crc kubenswrapper[4926]: I0312 18:27:01.225331 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" event={"ID":"6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69","Type":"ContainerStarted","Data":"f8918c17070f0ee95d203429bd314511a031eb1b8d8021bb41f8cd1c1487af96"} Mar 12 18:27:01 crc kubenswrapper[4926]: I0312 18:27:01.228464 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:27:01 crc kubenswrapper[4926]: I0312 18:27:01.259619 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" podStartSLOduration=3.25957685 podStartE2EDuration="3.25957685s" podCreationTimestamp="2026-03-12 18:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:27:01.2471572 +0000 UTC m=+1461.615783563" watchObservedRunningTime="2026-03-12 18:27:01.25957685 +0000 UTC m=+1461.628203183" Mar 12 18:27:08 crc kubenswrapper[4926]: I0312 18:27:08.761756 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5fc8b56cf-6cbwx" Mar 12 18:27:08 crc kubenswrapper[4926]: I0312 18:27:08.861485 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4ldnw"] Mar 12 18:27:08 crc kubenswrapper[4926]: I0312 18:27:08.861822 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" podUID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerName="dnsmasq-dns" containerID="cri-o://31005cd2c4f0038c1a2e2165529b7d457634779c21c8fccaed21553ed78e6d14" gracePeriod=10 Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.320539 4926 generic.go:334] "Generic (PLEG): container finished" podID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerID="31005cd2c4f0038c1a2e2165529b7d457634779c21c8fccaed21553ed78e6d14" exitCode=0 Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.320602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" event={"ID":"ce48eb28-755b-4a18-8bf1-212baff1f7fc","Type":"ContainerDied","Data":"31005cd2c4f0038c1a2e2165529b7d457634779c21c8fccaed21553ed78e6d14"} Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.462291 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.565976 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-nb\") pod \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.566091 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fvxr\" (UniqueName: \"kubernetes.io/projected/ce48eb28-755b-4a18-8bf1-212baff1f7fc-kube-api-access-5fvxr\") pod \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.566180 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-config\") pod \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.566259 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-openstack-edpm-ipam\") pod \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.566318 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-swift-storage-0\") pod \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.566393 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-sb\") pod \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.566425 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-svc\") pod \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\" (UID: \"ce48eb28-755b-4a18-8bf1-212baff1f7fc\") " Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.576510 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce48eb28-755b-4a18-8bf1-212baff1f7fc-kube-api-access-5fvxr" (OuterVolumeSpecName: "kube-api-access-5fvxr") pod "ce48eb28-755b-4a18-8bf1-212baff1f7fc" (UID: "ce48eb28-755b-4a18-8bf1-212baff1f7fc"). InnerVolumeSpecName "kube-api-access-5fvxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.622908 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ce48eb28-755b-4a18-8bf1-212baff1f7fc" (UID: "ce48eb28-755b-4a18-8bf1-212baff1f7fc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.626033 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce48eb28-755b-4a18-8bf1-212baff1f7fc" (UID: "ce48eb28-755b-4a18-8bf1-212baff1f7fc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.630206 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce48eb28-755b-4a18-8bf1-212baff1f7fc" (UID: "ce48eb28-755b-4a18-8bf1-212baff1f7fc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.631076 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce48eb28-755b-4a18-8bf1-212baff1f7fc" (UID: "ce48eb28-755b-4a18-8bf1-212baff1f7fc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.637086 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ce48eb28-755b-4a18-8bf1-212baff1f7fc" (UID: "ce48eb28-755b-4a18-8bf1-212baff1f7fc"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.644032 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-config" (OuterVolumeSpecName: "config") pod "ce48eb28-755b-4a18-8bf1-212baff1f7fc" (UID: "ce48eb28-755b-4a18-8bf1-212baff1f7fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.669826 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-config\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.669858 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.669871 4926 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.669881 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.669892 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.669902 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce48eb28-755b-4a18-8bf1-212baff1f7fc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:09 crc kubenswrapper[4926]: I0312 18:27:09.669912 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fvxr\" (UniqueName: \"kubernetes.io/projected/ce48eb28-755b-4a18-8bf1-212baff1f7fc-kube-api-access-5fvxr\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:10 crc kubenswrapper[4926]: I0312 18:27:10.337602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" event={"ID":"ce48eb28-755b-4a18-8bf1-212baff1f7fc","Type":"ContainerDied","Data":"73aeb25dea0852f705f0adfdaeaab49793c64b81144d0bee56d28bfe72283a50"} Mar 12 18:27:10 crc kubenswrapper[4926]: I0312 18:27:10.337685 4926 scope.go:117] "RemoveContainer" containerID="31005cd2c4f0038c1a2e2165529b7d457634779c21c8fccaed21553ed78e6d14" Mar 12 18:27:10 crc kubenswrapper[4926]: I0312 18:27:10.337693 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-4ldnw" Mar 12 18:27:10 crc kubenswrapper[4926]: I0312 18:27:10.388096 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4ldnw"] Mar 12 18:27:10 crc kubenswrapper[4926]: I0312 18:27:10.388728 4926 scope.go:117] "RemoveContainer" containerID="5c57deb5d6d68579f57781cc73b1f95fd91c467933afbbc36484489589a3e08d" Mar 12 18:27:10 crc kubenswrapper[4926]: I0312 18:27:10.405345 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-4ldnw"] Mar 12 18:27:10 crc kubenswrapper[4926]: I0312 18:27:10.504864 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" path="/var/lib/kubelet/pods/ce48eb28-755b-4a18-8bf1-212baff1f7fc/volumes" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.580082 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8cm8t"] Mar 12 18:27:13 crc kubenswrapper[4926]: E0312 18:27:13.581158 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerName="init" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.581175 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerName="init" Mar 12 18:27:13 crc kubenswrapper[4926]: E0312 18:27:13.581190 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerName="dnsmasq-dns" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.581196 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerName="dnsmasq-dns" Mar 12 18:27:13 crc kubenswrapper[4926]: E0312 18:27:13.581204 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerName="dnsmasq-dns" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.581211 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerName="dnsmasq-dns" Mar 12 18:27:13 crc kubenswrapper[4926]: E0312 18:27:13.581232 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerName="init" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.581238 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerName="init" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.581420 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d97cdd-8bee-43b1-a07c-fee61fceff3a" containerName="dnsmasq-dns" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.581453 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce48eb28-755b-4a18-8bf1-212baff1f7fc" containerName="dnsmasq-dns" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.582813 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.591345 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8cm8t"] Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.651535 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnlg2\" (UniqueName: \"kubernetes.io/projected/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-kube-api-access-bnlg2\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.651663 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-catalog-content\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.651730 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-utilities\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.755702 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnlg2\" (UniqueName: \"kubernetes.io/projected/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-kube-api-access-bnlg2\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.756029 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-catalog-content\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.756158 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-utilities\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.756894 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-utilities\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.757253 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-catalog-content\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.786580 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnlg2\" (UniqueName: \"kubernetes.io/projected/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-kube-api-access-bnlg2\") pod \"redhat-operators-8cm8t\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:13 crc kubenswrapper[4926]: I0312 18:27:13.965567 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.466411 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8cm8t"] Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.551413 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs"] Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.553713 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.555953 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.556039 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.556424 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.557003 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.591538 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs"] Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.674005 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fvjf\" (UniqueName: \"kubernetes.io/projected/58e2a40e-84ba-43fe-8087-ec2e917a2b34-kube-api-access-4fvjf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.674255 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.674298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.674466 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.776081 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.776552 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.776658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.776712 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fvjf\" (UniqueName: \"kubernetes.io/projected/58e2a40e-84ba-43fe-8087-ec2e917a2b34-kube-api-access-4fvjf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.784607 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.784921 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.785764 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.797877 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fvjf\" (UniqueName: \"kubernetes.io/projected/58e2a40e-84ba-43fe-8087-ec2e917a2b34-kube-api-access-4fvjf\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:14 crc kubenswrapper[4926]: I0312 18:27:14.976797 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:15 crc kubenswrapper[4926]: I0312 18:27:15.388199 4926 generic.go:334] "Generic (PLEG): container finished" podID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerID="c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045" exitCode=0 Mar 12 18:27:15 crc kubenswrapper[4926]: I0312 18:27:15.388361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cm8t" event={"ID":"f2f7ab38-c8b3-43a3-91f8-32168e7216ab","Type":"ContainerDied","Data":"c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045"} Mar 12 18:27:15 crc kubenswrapper[4926]: I0312 18:27:15.388505 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cm8t" event={"ID":"f2f7ab38-c8b3-43a3-91f8-32168e7216ab","Type":"ContainerStarted","Data":"f146c71e8411ea1922f76fefbfa5aefdbf611c51e0b51cca9e5da48e6e433b90"} Mar 12 18:27:15 crc kubenswrapper[4926]: I0312 18:27:15.559503 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs"] Mar 12 18:27:16 crc kubenswrapper[4926]: I0312 18:27:16.414003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" event={"ID":"58e2a40e-84ba-43fe-8087-ec2e917a2b34","Type":"ContainerStarted","Data":"79362af8345ee40914f1cad46dfe7258d5b7d2f58873fa86fb3735011e53e02e"} Mar 12 18:27:16 crc kubenswrapper[4926]: I0312 18:27:16.418176 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cm8t" event={"ID":"f2f7ab38-c8b3-43a3-91f8-32168e7216ab","Type":"ContainerStarted","Data":"6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124"} Mar 12 18:27:19 crc kubenswrapper[4926]: I0312 18:27:19.452494 4926 generic.go:334] "Generic (PLEG): container finished" podID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerID="6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124" exitCode=0 Mar 12 18:27:19 crc kubenswrapper[4926]: I0312 18:27:19.452551 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cm8t" event={"ID":"f2f7ab38-c8b3-43a3-91f8-32168e7216ab","Type":"ContainerDied","Data":"6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124"} Mar 12 18:27:20 crc kubenswrapper[4926]: I0312 18:27:20.464291 4926 generic.go:334] "Generic (PLEG): container finished" podID="9aba0434-585e-4355-8019-1612400b2350" containerID="f428e986bfed0b99e85dec3bb535e7ec0c4eb9551d8016425ac735d81340c54d" exitCode=0 Mar 12 18:27:20 crc kubenswrapper[4926]: I0312 18:27:20.464387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9aba0434-585e-4355-8019-1612400b2350","Type":"ContainerDied","Data":"f428e986bfed0b99e85dec3bb535e7ec0c4eb9551d8016425ac735d81340c54d"} Mar 12 18:27:21 crc kubenswrapper[4926]: I0312 18:27:21.482508 4926 generic.go:334] "Generic (PLEG): container finished" podID="b9b82f03-7ac1-4805-858b-708760b4e476" containerID="b958f13a715a00e9a33abc7ba17caf342fd31d1ad60a10155cff9bad1045c238" exitCode=0 Mar 12 18:27:21 crc kubenswrapper[4926]: I0312 18:27:21.482614 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b9b82f03-7ac1-4805-858b-708760b4e476","Type":"ContainerDied","Data":"b958f13a715a00e9a33abc7ba17caf342fd31d1ad60a10155cff9bad1045c238"} Mar 12 18:27:26 crc kubenswrapper[4926]: I0312 18:27:26.818371 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:27:26 crc kubenswrapper[4926]: I0312 18:27:26.821553 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.573815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b9b82f03-7ac1-4805-858b-708760b4e476","Type":"ContainerStarted","Data":"af39b4a41c7e394aacc8c9804a619a18aba2af552e6d25e9c1ebf02ad492191e"} Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.574517 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.577500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"9aba0434-585e-4355-8019-1612400b2350","Type":"ContainerStarted","Data":"dfba317cf8f9246b78f745195185ea9a285eabf0aed35addfc16d58bb52f783c"} Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.578011 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.580545 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" event={"ID":"58e2a40e-84ba-43fe-8087-ec2e917a2b34","Type":"ContainerStarted","Data":"e358207882b6752201641451ec266917c4f67e93fa12995e0e522f44e4c78575"} Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.583071 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cm8t" event={"ID":"f2f7ab38-c8b3-43a3-91f8-32168e7216ab","Type":"ContainerStarted","Data":"0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9"} Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.607133 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.607096246 podStartE2EDuration="42.607096246s" podCreationTimestamp="2026-03-12 18:26:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:27:27.596303938 +0000 UTC m=+1487.964930301" watchObservedRunningTime="2026-03-12 18:27:27.607096246 +0000 UTC m=+1487.975722579" Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.641827 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8cm8t" podStartSLOduration=3.702316326 podStartE2EDuration="14.641790215s" podCreationTimestamp="2026-03-12 18:27:13 +0000 UTC" firstStartedPulling="2026-03-12 18:27:15.390557515 +0000 UTC m=+1475.759183848" lastFinishedPulling="2026-03-12 18:27:26.330031404 +0000 UTC m=+1486.698657737" observedRunningTime="2026-03-12 18:27:27.62697491 +0000 UTC m=+1487.995601293" watchObservedRunningTime="2026-03-12 18:27:27.641790215 +0000 UTC m=+1488.010416548" Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.665357 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" podStartSLOduration=2.885864916 podStartE2EDuration="13.665327484s" podCreationTimestamp="2026-03-12 18:27:14 +0000 UTC" firstStartedPulling="2026-03-12 18:27:15.566792175 +0000 UTC m=+1475.935418508" lastFinishedPulling="2026-03-12 18:27:26.346254743 +0000 UTC m=+1486.714881076" observedRunningTime="2026-03-12 18:27:27.652490671 +0000 UTC m=+1488.021117014" watchObservedRunningTime="2026-03-12 18:27:27.665327484 +0000 UTC m=+1488.033953837" Mar 12 18:27:27 crc kubenswrapper[4926]: I0312 18:27:27.689123 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=45.689088109 podStartE2EDuration="45.689088109s" podCreationTimestamp="2026-03-12 18:26:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:27:27.679004873 +0000 UTC m=+1488.047631206" watchObservedRunningTime="2026-03-12 18:27:27.689088109 +0000 UTC m=+1488.057714452" Mar 12 18:27:33 crc kubenswrapper[4926]: I0312 18:27:33.965626 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:33 crc kubenswrapper[4926]: I0312 18:27:33.967852 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:27:35 crc kubenswrapper[4926]: I0312 18:27:35.015608 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8cm8t" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="registry-server" probeResult="failure" output=< Mar 12 18:27:35 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:27:35 crc kubenswrapper[4926]: > Mar 12 18:27:37 crc kubenswrapper[4926]: I0312 18:27:37.729238 4926 generic.go:334] "Generic (PLEG): container finished" podID="58e2a40e-84ba-43fe-8087-ec2e917a2b34" containerID="e358207882b6752201641451ec266917c4f67e93fa12995e0e522f44e4c78575" exitCode=0 Mar 12 18:27:37 crc kubenswrapper[4926]: I0312 18:27:37.729322 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" event={"ID":"58e2a40e-84ba-43fe-8087-ec2e917a2b34","Type":"ContainerDied","Data":"e358207882b6752201641451ec266917c4f67e93fa12995e0e522f44e4c78575"} Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.230854 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.322391 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-inventory\") pod \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.322809 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fvjf\" (UniqueName: \"kubernetes.io/projected/58e2a40e-84ba-43fe-8087-ec2e917a2b34-kube-api-access-4fvjf\") pod \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.322938 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-ssh-key-openstack-edpm-ipam\") pod \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.323223 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-repo-setup-combined-ca-bundle\") pod \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\" (UID: \"58e2a40e-84ba-43fe-8087-ec2e917a2b34\") " Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.329223 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e2a40e-84ba-43fe-8087-ec2e917a2b34-kube-api-access-4fvjf" (OuterVolumeSpecName: "kube-api-access-4fvjf") pod "58e2a40e-84ba-43fe-8087-ec2e917a2b34" (UID: "58e2a40e-84ba-43fe-8087-ec2e917a2b34"). InnerVolumeSpecName "kube-api-access-4fvjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.335707 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "58e2a40e-84ba-43fe-8087-ec2e917a2b34" (UID: "58e2a40e-84ba-43fe-8087-ec2e917a2b34"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.359310 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "58e2a40e-84ba-43fe-8087-ec2e917a2b34" (UID: "58e2a40e-84ba-43fe-8087-ec2e917a2b34"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.365382 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-inventory" (OuterVolumeSpecName: "inventory") pod "58e2a40e-84ba-43fe-8087-ec2e917a2b34" (UID: "58e2a40e-84ba-43fe-8087-ec2e917a2b34"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.425587 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.425626 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fvjf\" (UniqueName: \"kubernetes.io/projected/58e2a40e-84ba-43fe-8087-ec2e917a2b34-kube-api-access-4fvjf\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.425639 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.425648 4926 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e2a40e-84ba-43fe-8087-ec2e917a2b34-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.753318 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" event={"ID":"58e2a40e-84ba-43fe-8087-ec2e917a2b34","Type":"ContainerDied","Data":"79362af8345ee40914f1cad46dfe7258d5b7d2f58873fa86fb3735011e53e02e"} Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.753662 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79362af8345ee40914f1cad46dfe7258d5b7d2f58873fa86fb3735011e53e02e" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.753369 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.850885 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z"] Mar 12 18:27:39 crc kubenswrapper[4926]: E0312 18:27:39.851294 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58e2a40e-84ba-43fe-8087-ec2e917a2b34" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.851329 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="58e2a40e-84ba-43fe-8087-ec2e917a2b34" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.851592 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="58e2a40e-84ba-43fe-8087-ec2e917a2b34" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.852210 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.855966 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.856121 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.856379 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.856991 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.866110 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z"] Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.935588 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.935778 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.936184 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/75a3208b-42f5-412e-a503-ac328f7d9967-kube-api-access-mpczd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:39 crc kubenswrapper[4926]: I0312 18:27:39.936341 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.038539 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/75a3208b-42f5-412e-a503-ac328f7d9967-kube-api-access-mpczd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.038632 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.038719 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.038796 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.043620 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.044053 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.051218 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.066024 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/75a3208b-42f5-412e-a503-ac328f7d9967-kube-api-access-mpczd\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.175732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.763117 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z"] Mar 12 18:27:40 crc kubenswrapper[4926]: I0312 18:27:40.774913 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:27:41 crc kubenswrapper[4926]: I0312 18:27:41.246198 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:27:41 crc kubenswrapper[4926]: I0312 18:27:41.789404 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" event={"ID":"75a3208b-42f5-412e-a503-ac328f7d9967","Type":"ContainerStarted","Data":"6e0e0f76508b934742dbabb4b94eeab9a4fa625bd527c90d90fa0cda46ddea4a"} Mar 12 18:27:41 crc kubenswrapper[4926]: I0312 18:27:41.789818 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" event={"ID":"75a3208b-42f5-412e-a503-ac328f7d9967","Type":"ContainerStarted","Data":"beacf7f2c4ecf74d1f45669096c28cb46b36b877d88778fb003f1ce1efb01ae4"} Mar 12 18:27:41 crc kubenswrapper[4926]: I0312 18:27:41.817019 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" podStartSLOduration=2.348196392 podStartE2EDuration="2.81699493s" podCreationTimestamp="2026-03-12 18:27:39 +0000 UTC" firstStartedPulling="2026-03-12 18:27:40.774729893 +0000 UTC m=+1501.143356226" lastFinishedPulling="2026-03-12 18:27:41.243528421 +0000 UTC m=+1501.612154764" observedRunningTime="2026-03-12 18:27:41.807856718 +0000 UTC m=+1502.176483141" watchObservedRunningTime="2026-03-12 18:27:41.81699493 +0000 UTC m=+1502.185621263" Mar 12 18:27:43 crc kubenswrapper[4926]: I0312 18:27:43.390689 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 12 18:27:45 crc kubenswrapper[4926]: I0312 18:27:45.007666 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8cm8t" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="registry-server" probeResult="failure" output=< Mar 12 18:27:45 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:27:45 crc kubenswrapper[4926]: > Mar 12 18:27:45 crc kubenswrapper[4926]: I0312 18:27:45.720641 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 12 18:27:55 crc kubenswrapper[4926]: I0312 18:27:55.010773 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8cm8t" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="registry-server" probeResult="failure" output=< Mar 12 18:27:55 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:27:55 crc kubenswrapper[4926]: > Mar 12 18:27:56 crc kubenswrapper[4926]: I0312 18:27:56.817472 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:27:56 crc kubenswrapper[4926]: I0312 18:27:56.817779 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.145093 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555668-xnfj4"] Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.146925 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555668-xnfj4" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.150816 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.151019 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.152593 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.170880 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555668-xnfj4"] Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.308178 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjngx\" (UniqueName: \"kubernetes.io/projected/ddbd98cc-c135-485c-9b18-c31f19321a38-kube-api-access-xjngx\") pod \"auto-csr-approver-29555668-xnfj4\" (UID: \"ddbd98cc-c135-485c-9b18-c31f19321a38\") " pod="openshift-infra/auto-csr-approver-29555668-xnfj4" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.410133 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjngx\" (UniqueName: \"kubernetes.io/projected/ddbd98cc-c135-485c-9b18-c31f19321a38-kube-api-access-xjngx\") pod \"auto-csr-approver-29555668-xnfj4\" (UID: \"ddbd98cc-c135-485c-9b18-c31f19321a38\") " pod="openshift-infra/auto-csr-approver-29555668-xnfj4" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.454053 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjngx\" (UniqueName: \"kubernetes.io/projected/ddbd98cc-c135-485c-9b18-c31f19321a38-kube-api-access-xjngx\") pod \"auto-csr-approver-29555668-xnfj4\" (UID: \"ddbd98cc-c135-485c-9b18-c31f19321a38\") " pod="openshift-infra/auto-csr-approver-29555668-xnfj4" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.470073 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555668-xnfj4" Mar 12 18:28:00 crc kubenswrapper[4926]: I0312 18:28:00.966286 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555668-xnfj4"] Mar 12 18:28:00 crc kubenswrapper[4926]: W0312 18:28:00.982334 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddbd98cc_c135_485c_9b18_c31f19321a38.slice/crio-0dc8726e4b9c908068de0db72c97f99f237cbd253cb409610e6f712f3dbcd160 WatchSource:0}: Error finding container 0dc8726e4b9c908068de0db72c97f99f237cbd253cb409610e6f712f3dbcd160: Status 404 returned error can't find the container with id 0dc8726e4b9c908068de0db72c97f99f237cbd253cb409610e6f712f3dbcd160 Mar 12 18:28:01 crc kubenswrapper[4926]: I0312 18:28:01.015555 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555668-xnfj4" event={"ID":"ddbd98cc-c135-485c-9b18-c31f19321a38","Type":"ContainerStarted","Data":"0dc8726e4b9c908068de0db72c97f99f237cbd253cb409610e6f712f3dbcd160"} Mar 12 18:28:03 crc kubenswrapper[4926]: I0312 18:28:03.037201 4926 generic.go:334] "Generic (PLEG): container finished" podID="ddbd98cc-c135-485c-9b18-c31f19321a38" containerID="74c22411866d45d292c4082e0d4be48c52ad4683049047a45e2de32d9da3478f" exitCode=0 Mar 12 18:28:03 crc kubenswrapper[4926]: I0312 18:28:03.037294 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555668-xnfj4" event={"ID":"ddbd98cc-c135-485c-9b18-c31f19321a38","Type":"ContainerDied","Data":"74c22411866d45d292c4082e0d4be48c52ad4683049047a45e2de32d9da3478f"} Mar 12 18:28:04 crc kubenswrapper[4926]: I0312 18:28:04.037381 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:28:04 crc kubenswrapper[4926]: I0312 18:28:04.092027 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:28:04 crc kubenswrapper[4926]: I0312 18:28:04.280547 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8cm8t"] Mar 12 18:28:04 crc kubenswrapper[4926]: I0312 18:28:04.377063 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555668-xnfj4" Mar 12 18:28:04 crc kubenswrapper[4926]: I0312 18:28:04.506028 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjngx\" (UniqueName: \"kubernetes.io/projected/ddbd98cc-c135-485c-9b18-c31f19321a38-kube-api-access-xjngx\") pod \"ddbd98cc-c135-485c-9b18-c31f19321a38\" (UID: \"ddbd98cc-c135-485c-9b18-c31f19321a38\") " Mar 12 18:28:04 crc kubenswrapper[4926]: I0312 18:28:04.522313 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddbd98cc-c135-485c-9b18-c31f19321a38-kube-api-access-xjngx" (OuterVolumeSpecName: "kube-api-access-xjngx") pod "ddbd98cc-c135-485c-9b18-c31f19321a38" (UID: "ddbd98cc-c135-485c-9b18-c31f19321a38"). InnerVolumeSpecName "kube-api-access-xjngx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:28:04 crc kubenswrapper[4926]: I0312 18:28:04.608632 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjngx\" (UniqueName: \"kubernetes.io/projected/ddbd98cc-c135-485c-9b18-c31f19321a38-kube-api-access-xjngx\") on node \"crc\" DevicePath \"\"" Mar 12 18:28:05 crc kubenswrapper[4926]: I0312 18:28:05.066499 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555668-xnfj4" event={"ID":"ddbd98cc-c135-485c-9b18-c31f19321a38","Type":"ContainerDied","Data":"0dc8726e4b9c908068de0db72c97f99f237cbd253cb409610e6f712f3dbcd160"} Mar 12 18:28:05 crc kubenswrapper[4926]: I0312 18:28:05.066533 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555668-xnfj4" Mar 12 18:28:05 crc kubenswrapper[4926]: I0312 18:28:05.066572 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0dc8726e4b9c908068de0db72c97f99f237cbd253cb409610e6f712f3dbcd160" Mar 12 18:28:05 crc kubenswrapper[4926]: E0312 18:28:05.262740 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddbd98cc_c135_485c_9b18_c31f19321a38.slice\": RecentStats: unable to find data in memory cache]" Mar 12 18:28:05 crc kubenswrapper[4926]: I0312 18:28:05.446420 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555662-wmwgq"] Mar 12 18:28:05 crc kubenswrapper[4926]: I0312 18:28:05.453522 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555662-wmwgq"] Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.076673 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8cm8t" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="registry-server" containerID="cri-o://0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9" gracePeriod=2 Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.502527 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3def1951-b6f9-4621-8428-b3e169e34279" path="/var/lib/kubelet/pods/3def1951-b6f9-4621-8428-b3e169e34279/volumes" Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.633788 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.749685 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-catalog-content\") pod \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.749799 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-utilities\") pod \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.749967 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnlg2\" (UniqueName: \"kubernetes.io/projected/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-kube-api-access-bnlg2\") pod \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\" (UID: \"f2f7ab38-c8b3-43a3-91f8-32168e7216ab\") " Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.750514 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-utilities" (OuterVolumeSpecName: "utilities") pod "f2f7ab38-c8b3-43a3-91f8-32168e7216ab" (UID: "f2f7ab38-c8b3-43a3-91f8-32168e7216ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.759281 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-kube-api-access-bnlg2" (OuterVolumeSpecName: "kube-api-access-bnlg2") pod "f2f7ab38-c8b3-43a3-91f8-32168e7216ab" (UID: "f2f7ab38-c8b3-43a3-91f8-32168e7216ab"). InnerVolumeSpecName "kube-api-access-bnlg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.852386 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.852425 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnlg2\" (UniqueName: \"kubernetes.io/projected/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-kube-api-access-bnlg2\") on node \"crc\" DevicePath \"\"" Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.878877 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2f7ab38-c8b3-43a3-91f8-32168e7216ab" (UID: "f2f7ab38-c8b3-43a3-91f8-32168e7216ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:28:06 crc kubenswrapper[4926]: I0312 18:28:06.954535 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2f7ab38-c8b3-43a3-91f8-32168e7216ab-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.092283 4926 generic.go:334] "Generic (PLEG): container finished" podID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerID="0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9" exitCode=0 Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.092357 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cm8t" event={"ID":"f2f7ab38-c8b3-43a3-91f8-32168e7216ab","Type":"ContainerDied","Data":"0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9"} Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.092409 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8cm8t" event={"ID":"f2f7ab38-c8b3-43a3-91f8-32168e7216ab","Type":"ContainerDied","Data":"f146c71e8411ea1922f76fefbfa5aefdbf611c51e0b51cca9e5da48e6e433b90"} Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.092480 4926 scope.go:117] "RemoveContainer" containerID="0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.093683 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8cm8t" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.118662 4926 scope.go:117] "RemoveContainer" containerID="6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.157983 4926 scope.go:117] "RemoveContainer" containerID="c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.162582 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8cm8t"] Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.179233 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8cm8t"] Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.192705 4926 scope.go:117] "RemoveContainer" containerID="0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9" Mar 12 18:28:07 crc kubenswrapper[4926]: E0312 18:28:07.193206 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9\": container with ID starting with 0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9 not found: ID does not exist" containerID="0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.193249 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9"} err="failed to get container status \"0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9\": rpc error: code = NotFound desc = could not find container \"0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9\": container with ID starting with 0a9c576f66c558f54f81f42fdf69438a16dadc2a363b3f13d9021793b3966fb9 not found: ID does not exist" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.193278 4926 scope.go:117] "RemoveContainer" containerID="6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124" Mar 12 18:28:07 crc kubenswrapper[4926]: E0312 18:28:07.193582 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124\": container with ID starting with 6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124 not found: ID does not exist" containerID="6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.193608 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124"} err="failed to get container status \"6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124\": rpc error: code = NotFound desc = could not find container \"6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124\": container with ID starting with 6c8be519f6eb6a16345af9a869dccf59f8f37215a19f757ecf6f68becb437124 not found: ID does not exist" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.193622 4926 scope.go:117] "RemoveContainer" containerID="c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045" Mar 12 18:28:07 crc kubenswrapper[4926]: E0312 18:28:07.193908 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045\": container with ID starting with c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045 not found: ID does not exist" containerID="c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045" Mar 12 18:28:07 crc kubenswrapper[4926]: I0312 18:28:07.193932 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045"} err="failed to get container status \"c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045\": rpc error: code = NotFound desc = could not find container \"c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045\": container with ID starting with c7612d6f6ac3e40212f1f47065b076937ee6ab26f5cab2dec5bc30b2d4230045 not found: ID does not exist" Mar 12 18:28:08 crc kubenswrapper[4926]: I0312 18:28:08.517754 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" path="/var/lib/kubelet/pods/f2f7ab38-c8b3-43a3-91f8-32168e7216ab/volumes" Mar 12 18:28:14 crc kubenswrapper[4926]: I0312 18:28:14.643078 4926 scope.go:117] "RemoveContainer" containerID="6e11e8ecbd719cce51e25090959e2c12962023683d089714c24c4c379fbeaa75" Mar 12 18:28:14 crc kubenswrapper[4926]: I0312 18:28:14.695475 4926 scope.go:117] "RemoveContainer" containerID="bf4088b7010b072d07c97127da27603d49f76d54fb965fcf30cdce478bed5715" Mar 12 18:28:26 crc kubenswrapper[4926]: I0312 18:28:26.817222 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:28:26 crc kubenswrapper[4926]: I0312 18:28:26.818682 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:28:26 crc kubenswrapper[4926]: I0312 18:28:26.818756 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:28:26 crc kubenswrapper[4926]: I0312 18:28:26.819506 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"759fd18072cdf8fcc7bc2d92cc950b5720a437d7e4487f5098fffd2244e21cde"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:28:26 crc kubenswrapper[4926]: I0312 18:28:26.819567 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://759fd18072cdf8fcc7bc2d92cc950b5720a437d7e4487f5098fffd2244e21cde" gracePeriod=600 Mar 12 18:28:27 crc kubenswrapper[4926]: I0312 18:28:27.312114 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="759fd18072cdf8fcc7bc2d92cc950b5720a437d7e4487f5098fffd2244e21cde" exitCode=0 Mar 12 18:28:27 crc kubenswrapper[4926]: I0312 18:28:27.312189 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"759fd18072cdf8fcc7bc2d92cc950b5720a437d7e4487f5098fffd2244e21cde"} Mar 12 18:28:27 crc kubenswrapper[4926]: I0312 18:28:27.312505 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e"} Mar 12 18:28:27 crc kubenswrapper[4926]: I0312 18:28:27.312529 4926 scope.go:117] "RemoveContainer" containerID="9728bd6132bdd9ab31a71d0a44779a02f515c39e712bb7cc4f8a85610efe739f" Mar 12 18:29:14 crc kubenswrapper[4926]: I0312 18:29:14.828804 4926 scope.go:117] "RemoveContainer" containerID="334e13044642c4b0a613f2bfccdc8e4e0067f4f734093e62a567052168539e3c" Mar 12 18:29:14 crc kubenswrapper[4926]: I0312 18:29:14.880593 4926 scope.go:117] "RemoveContainer" containerID="f80e900658b9e00858dd34680664f8b822bb4e0eba4897392ac1a1da7732a56c" Mar 12 18:29:14 crc kubenswrapper[4926]: I0312 18:29:14.931968 4926 scope.go:117] "RemoveContainer" containerID="f2d7197b9b3bf88ce380eff3700bdf47487a3209f60cede9380f4a5b74fd2a07" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.629416 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-97nz8"] Mar 12 18:29:34 crc kubenswrapper[4926]: E0312 18:29:34.630330 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="registry-server" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.630342 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="registry-server" Mar 12 18:29:34 crc kubenswrapper[4926]: E0312 18:29:34.630356 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="extract-utilities" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.630362 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="extract-utilities" Mar 12 18:29:34 crc kubenswrapper[4926]: E0312 18:29:34.630370 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="extract-content" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.630376 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="extract-content" Mar 12 18:29:34 crc kubenswrapper[4926]: E0312 18:29:34.630412 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddbd98cc-c135-485c-9b18-c31f19321a38" containerName="oc" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.630418 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddbd98cc-c135-485c-9b18-c31f19321a38" containerName="oc" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.630606 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2f7ab38-c8b3-43a3-91f8-32168e7216ab" containerName="registry-server" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.630624 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddbd98cc-c135-485c-9b18-c31f19321a38" containerName="oc" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.632670 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.654838 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-97nz8"] Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.728638 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-utilities\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.728991 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcggp\" (UniqueName: \"kubernetes.io/projected/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-kube-api-access-gcggp\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.729071 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-catalog-content\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.830560 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-utilities\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.830675 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcggp\" (UniqueName: \"kubernetes.io/projected/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-kube-api-access-gcggp\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.830709 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-catalog-content\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.831333 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-catalog-content\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.831348 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-utilities\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.852460 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcggp\" (UniqueName: \"kubernetes.io/projected/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-kube-api-access-gcggp\") pod \"certified-operators-97nz8\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:34 crc kubenswrapper[4926]: I0312 18:29:34.961046 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:35 crc kubenswrapper[4926]: I0312 18:29:35.500802 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-97nz8"] Mar 12 18:29:35 crc kubenswrapper[4926]: W0312 18:29:35.505899 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa4d5480_0ddc_4ef0_8c78_2dc18fedd103.slice/crio-37034daa38cf52c12ba3c600dedf2eb059e0d574ad95211676b768bc7e090443 WatchSource:0}: Error finding container 37034daa38cf52c12ba3c600dedf2eb059e0d574ad95211676b768bc7e090443: Status 404 returned error can't find the container with id 37034daa38cf52c12ba3c600dedf2eb059e0d574ad95211676b768bc7e090443 Mar 12 18:29:36 crc kubenswrapper[4926]: I0312 18:29:36.068375 4926 generic.go:334] "Generic (PLEG): container finished" podID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerID="7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1" exitCode=0 Mar 12 18:29:36 crc kubenswrapper[4926]: I0312 18:29:36.068471 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97nz8" event={"ID":"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103","Type":"ContainerDied","Data":"7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1"} Mar 12 18:29:36 crc kubenswrapper[4926]: I0312 18:29:36.068741 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97nz8" event={"ID":"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103","Type":"ContainerStarted","Data":"37034daa38cf52c12ba3c600dedf2eb059e0d574ad95211676b768bc7e090443"} Mar 12 18:29:37 crc kubenswrapper[4926]: I0312 18:29:37.079598 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97nz8" event={"ID":"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103","Type":"ContainerStarted","Data":"484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a"} Mar 12 18:29:38 crc kubenswrapper[4926]: I0312 18:29:38.092805 4926 generic.go:334] "Generic (PLEG): container finished" podID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerID="484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a" exitCode=0 Mar 12 18:29:38 crc kubenswrapper[4926]: I0312 18:29:38.092855 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97nz8" event={"ID":"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103","Type":"ContainerDied","Data":"484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a"} Mar 12 18:29:39 crc kubenswrapper[4926]: I0312 18:29:39.107234 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97nz8" event={"ID":"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103","Type":"ContainerStarted","Data":"af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7"} Mar 12 18:29:39 crc kubenswrapper[4926]: I0312 18:29:39.128508 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-97nz8" podStartSLOduration=2.418181901 podStartE2EDuration="5.12848204s" podCreationTimestamp="2026-03-12 18:29:34 +0000 UTC" firstStartedPulling="2026-03-12 18:29:36.070171425 +0000 UTC m=+1616.438797758" lastFinishedPulling="2026-03-12 18:29:38.780471564 +0000 UTC m=+1619.149097897" observedRunningTime="2026-03-12 18:29:39.128363927 +0000 UTC m=+1619.496990310" watchObservedRunningTime="2026-03-12 18:29:39.12848204 +0000 UTC m=+1619.497108373" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.014370 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5k9gj"] Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.016578 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.029719 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5k9gj"] Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.141157 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0817ed2e-acdd-41b8-b210-84281525839f-catalog-content\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.141203 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbjqw\" (UniqueName: \"kubernetes.io/projected/0817ed2e-acdd-41b8-b210-84281525839f-kube-api-access-lbjqw\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.141331 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0817ed2e-acdd-41b8-b210-84281525839f-utilities\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.243314 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0817ed2e-acdd-41b8-b210-84281525839f-utilities\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.243405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0817ed2e-acdd-41b8-b210-84281525839f-catalog-content\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.243429 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbjqw\" (UniqueName: \"kubernetes.io/projected/0817ed2e-acdd-41b8-b210-84281525839f-kube-api-access-lbjqw\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.243800 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0817ed2e-acdd-41b8-b210-84281525839f-utilities\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.244169 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0817ed2e-acdd-41b8-b210-84281525839f-catalog-content\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.265736 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbjqw\" (UniqueName: \"kubernetes.io/projected/0817ed2e-acdd-41b8-b210-84281525839f-kube-api-access-lbjqw\") pod \"community-operators-5k9gj\" (UID: \"0817ed2e-acdd-41b8-b210-84281525839f\") " pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.335259 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:40 crc kubenswrapper[4926]: I0312 18:29:40.690379 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5k9gj"] Mar 12 18:29:41 crc kubenswrapper[4926]: I0312 18:29:41.132056 4926 generic.go:334] "Generic (PLEG): container finished" podID="0817ed2e-acdd-41b8-b210-84281525839f" containerID="8e93fac4e2765b9277b5b265a7c08055ab7f2b0fb83f9631b99c08b9f04874cf" exitCode=0 Mar 12 18:29:41 crc kubenswrapper[4926]: I0312 18:29:41.132114 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5k9gj" event={"ID":"0817ed2e-acdd-41b8-b210-84281525839f","Type":"ContainerDied","Data":"8e93fac4e2765b9277b5b265a7c08055ab7f2b0fb83f9631b99c08b9f04874cf"} Mar 12 18:29:41 crc kubenswrapper[4926]: I0312 18:29:41.132174 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5k9gj" event={"ID":"0817ed2e-acdd-41b8-b210-84281525839f","Type":"ContainerStarted","Data":"838aa39e9c3fa588ebc1ec1db1d9fae71627a58effba2a45c640d72c9ec4e6aa"} Mar 12 18:29:44 crc kubenswrapper[4926]: I0312 18:29:44.962153 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:44 crc kubenswrapper[4926]: I0312 18:29:44.962659 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:45 crc kubenswrapper[4926]: I0312 18:29:45.006721 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:45 crc kubenswrapper[4926]: I0312 18:29:45.231977 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:45 crc kubenswrapper[4926]: I0312 18:29:45.291616 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-97nz8"] Mar 12 18:29:46 crc kubenswrapper[4926]: I0312 18:29:46.177866 4926 generic.go:334] "Generic (PLEG): container finished" podID="0817ed2e-acdd-41b8-b210-84281525839f" containerID="9f135147b5771f7de5d2ccfbfc2d02a3d2a2771ee65bbf1e8e1f0a86854a1e59" exitCode=0 Mar 12 18:29:46 crc kubenswrapper[4926]: I0312 18:29:46.177937 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5k9gj" event={"ID":"0817ed2e-acdd-41b8-b210-84281525839f","Type":"ContainerDied","Data":"9f135147b5771f7de5d2ccfbfc2d02a3d2a2771ee65bbf1e8e1f0a86854a1e59"} Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.191630 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-97nz8" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="registry-server" containerID="cri-o://af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7" gracePeriod=2 Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.192163 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5k9gj" event={"ID":"0817ed2e-acdd-41b8-b210-84281525839f","Type":"ContainerStarted","Data":"7abe901bddf6d43393bae272cbfbd5c3f58b73de8b29e2c5d9741cbc83e0fbc0"} Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.240582 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5k9gj" podStartSLOduration=2.802052173 podStartE2EDuration="8.240560076s" podCreationTimestamp="2026-03-12 18:29:39 +0000 UTC" firstStartedPulling="2026-03-12 18:29:41.13476528 +0000 UTC m=+1621.503391613" lastFinishedPulling="2026-03-12 18:29:46.573273143 +0000 UTC m=+1626.941899516" observedRunningTime="2026-03-12 18:29:47.211507974 +0000 UTC m=+1627.580134337" watchObservedRunningTime="2026-03-12 18:29:47.240560076 +0000 UTC m=+1627.609186409" Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.665673 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.816688 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-catalog-content\") pod \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.817074 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-utilities\") pod \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.817317 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcggp\" (UniqueName: \"kubernetes.io/projected/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-kube-api-access-gcggp\") pod \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\" (UID: \"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103\") " Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.817916 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-utilities" (OuterVolumeSpecName: "utilities") pod "aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" (UID: "aa4d5480-0ddc-4ef0-8c78-2dc18fedd103"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.823124 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-kube-api-access-gcggp" (OuterVolumeSpecName: "kube-api-access-gcggp") pod "aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" (UID: "aa4d5480-0ddc-4ef0-8c78-2dc18fedd103"). InnerVolumeSpecName "kube-api-access-gcggp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.920660 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.920705 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcggp\" (UniqueName: \"kubernetes.io/projected/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-kube-api-access-gcggp\") on node \"crc\" DevicePath \"\"" Mar 12 18:29:47 crc kubenswrapper[4926]: I0312 18:29:47.992399 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" (UID: "aa4d5480-0ddc-4ef0-8c78-2dc18fedd103"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.023215 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.210174 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-97nz8" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.210196 4926 generic.go:334] "Generic (PLEG): container finished" podID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerID="af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7" exitCode=0 Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.210247 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97nz8" event={"ID":"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103","Type":"ContainerDied","Data":"af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7"} Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.211518 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-97nz8" event={"ID":"aa4d5480-0ddc-4ef0-8c78-2dc18fedd103","Type":"ContainerDied","Data":"37034daa38cf52c12ba3c600dedf2eb059e0d574ad95211676b768bc7e090443"} Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.211543 4926 scope.go:117] "RemoveContainer" containerID="af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.242591 4926 scope.go:117] "RemoveContainer" containerID="484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.258368 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-97nz8"] Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.269877 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-97nz8"] Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.284778 4926 scope.go:117] "RemoveContainer" containerID="7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.330617 4926 scope.go:117] "RemoveContainer" containerID="af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7" Mar 12 18:29:48 crc kubenswrapper[4926]: E0312 18:29:48.333245 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7\": container with ID starting with af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7 not found: ID does not exist" containerID="af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.333291 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7"} err="failed to get container status \"af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7\": rpc error: code = NotFound desc = could not find container \"af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7\": container with ID starting with af770418be95ad042053034dde6f8e0b355bb6b982997db4c758ea7364966da7 not found: ID does not exist" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.333319 4926 scope.go:117] "RemoveContainer" containerID="484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a" Mar 12 18:29:48 crc kubenswrapper[4926]: E0312 18:29:48.333655 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a\": container with ID starting with 484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a not found: ID does not exist" containerID="484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.334563 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a"} err="failed to get container status \"484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a\": rpc error: code = NotFound desc = could not find container \"484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a\": container with ID starting with 484f44600934344af2c346f90fc870cde957b1b9ae36fc19a8cb8fd79089750a not found: ID does not exist" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.334598 4926 scope.go:117] "RemoveContainer" containerID="7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1" Mar 12 18:29:48 crc kubenswrapper[4926]: E0312 18:29:48.334894 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1\": container with ID starting with 7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1 not found: ID does not exist" containerID="7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.334923 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1"} err="failed to get container status \"7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1\": rpc error: code = NotFound desc = could not find container \"7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1\": container with ID starting with 7d5874fb1e97421d1554dc44919f59f132287bb2c204fb0390ffc107768b89d1 not found: ID does not exist" Mar 12 18:29:48 crc kubenswrapper[4926]: I0312 18:29:48.503985 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" path="/var/lib/kubelet/pods/aa4d5480-0ddc-4ef0-8c78-2dc18fedd103/volumes" Mar 12 18:29:50 crc kubenswrapper[4926]: I0312 18:29:50.335525 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:50 crc kubenswrapper[4926]: I0312 18:29:50.335966 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:50 crc kubenswrapper[4926]: I0312 18:29:50.382573 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:51 crc kubenswrapper[4926]: I0312 18:29:51.310673 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5k9gj" Mar 12 18:29:51 crc kubenswrapper[4926]: I0312 18:29:51.493937 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5k9gj"] Mar 12 18:29:51 crc kubenswrapper[4926]: I0312 18:29:51.653123 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-89rth"] Mar 12 18:29:51 crc kubenswrapper[4926]: I0312 18:29:51.653422 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-89rth" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="registry-server" containerID="cri-o://b6bb1201cf8704c1c5218ae5ad6eb53e93f35257e13f18d1b638bfe3b2da6322" gracePeriod=2 Mar 12 18:29:52 crc kubenswrapper[4926]: I0312 18:29:52.266177 4926 generic.go:334] "Generic (PLEG): container finished" podID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerID="b6bb1201cf8704c1c5218ae5ad6eb53e93f35257e13f18d1b638bfe3b2da6322" exitCode=0 Mar 12 18:29:52 crc kubenswrapper[4926]: I0312 18:29:52.266469 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89rth" event={"ID":"8ed429f6-2923-42c2-a3b5-402c3dff4858","Type":"ContainerDied","Data":"b6bb1201cf8704c1c5218ae5ad6eb53e93f35257e13f18d1b638bfe3b2da6322"} Mar 12 18:29:52 crc kubenswrapper[4926]: I0312 18:29:52.907171 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.020561 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-catalog-content\") pod \"8ed429f6-2923-42c2-a3b5-402c3dff4858\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.020733 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-utilities\") pod \"8ed429f6-2923-42c2-a3b5-402c3dff4858\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.020844 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5zfs\" (UniqueName: \"kubernetes.io/projected/8ed429f6-2923-42c2-a3b5-402c3dff4858-kube-api-access-c5zfs\") pod \"8ed429f6-2923-42c2-a3b5-402c3dff4858\" (UID: \"8ed429f6-2923-42c2-a3b5-402c3dff4858\") " Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.022222 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-utilities" (OuterVolumeSpecName: "utilities") pod "8ed429f6-2923-42c2-a3b5-402c3dff4858" (UID: "8ed429f6-2923-42c2-a3b5-402c3dff4858"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.029564 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ed429f6-2923-42c2-a3b5-402c3dff4858-kube-api-access-c5zfs" (OuterVolumeSpecName: "kube-api-access-c5zfs") pod "8ed429f6-2923-42c2-a3b5-402c3dff4858" (UID: "8ed429f6-2923-42c2-a3b5-402c3dff4858"). InnerVolumeSpecName "kube-api-access-c5zfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.090265 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8ed429f6-2923-42c2-a3b5-402c3dff4858" (UID: "8ed429f6-2923-42c2-a3b5-402c3dff4858"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.123265 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.123296 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8ed429f6-2923-42c2-a3b5-402c3dff4858-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.123307 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5zfs\" (UniqueName: \"kubernetes.io/projected/8ed429f6-2923-42c2-a3b5-402c3dff4858-kube-api-access-c5zfs\") on node \"crc\" DevicePath \"\"" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.278547 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-89rth" event={"ID":"8ed429f6-2923-42c2-a3b5-402c3dff4858","Type":"ContainerDied","Data":"75b5fa3c244e1a8120e460f27209d8d4bc13d27e6ea5e2efce3f0acb11f95356"} Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.278587 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-89rth" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.278616 4926 scope.go:117] "RemoveContainer" containerID="b6bb1201cf8704c1c5218ae5ad6eb53e93f35257e13f18d1b638bfe3b2da6322" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.302605 4926 scope.go:117] "RemoveContainer" containerID="2efb7a52879f66525ffb82d5cb820bfe29543aedd6332a889b48b179b150edf8" Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.315692 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-89rth"] Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.328595 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-89rth"] Mar 12 18:29:53 crc kubenswrapper[4926]: I0312 18:29:53.333787 4926 scope.go:117] "RemoveContainer" containerID="b4a2b0d2b334eaa88bb6c32b357510979e7382c6582dd687f2155b5f20b07067" Mar 12 18:29:54 crc kubenswrapper[4926]: I0312 18:29:54.507499 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" path="/var/lib/kubelet/pods/8ed429f6-2923-42c2-a3b5-402c3dff4858/volumes" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.062251 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-29mgs"] Mar 12 18:29:59 crc kubenswrapper[4926]: E0312 18:29:59.063981 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="registry-server" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064004 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="registry-server" Mar 12 18:29:59 crc kubenswrapper[4926]: E0312 18:29:59.064021 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="extract-content" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064027 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="extract-content" Mar 12 18:29:59 crc kubenswrapper[4926]: E0312 18:29:59.064075 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="extract-content" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064082 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="extract-content" Mar 12 18:29:59 crc kubenswrapper[4926]: E0312 18:29:59.064106 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="extract-utilities" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064116 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="extract-utilities" Mar 12 18:29:59 crc kubenswrapper[4926]: E0312 18:29:59.064131 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="registry-server" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064139 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="registry-server" Mar 12 18:29:59 crc kubenswrapper[4926]: E0312 18:29:59.064154 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="extract-utilities" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064162 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="extract-utilities" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064425 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa4d5480-0ddc-4ef0-8c78-2dc18fedd103" containerName="registry-server" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.064479 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed429f6-2923-42c2-a3b5-402c3dff4858" containerName="registry-server" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.066317 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.090575 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29mgs"] Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.265781 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g5dw\" (UniqueName: \"kubernetes.io/projected/86dc3218-ffd0-4d12-af08-2c1323f6b670-kube-api-access-5g5dw\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.265981 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-catalog-content\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.266076 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-utilities\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.368112 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-catalog-content\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.368218 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-utilities\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.368271 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g5dw\" (UniqueName: \"kubernetes.io/projected/86dc3218-ffd0-4d12-af08-2c1323f6b670-kube-api-access-5g5dw\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.368744 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-catalog-content\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.368782 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-utilities\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.391348 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g5dw\" (UniqueName: \"kubernetes.io/projected/86dc3218-ffd0-4d12-af08-2c1323f6b670-kube-api-access-5g5dw\") pod \"redhat-marketplace-29mgs\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.392164 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:29:59 crc kubenswrapper[4926]: I0312 18:29:59.920652 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29mgs"] Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.171429 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555670-v8lqv"] Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.173847 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555670-v8lqv" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.185169 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx"] Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.186518 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.186660 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.186796 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.186796 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.187922 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxtgk\" (UniqueName: \"kubernetes.io/projected/daa768ee-1644-40fa-8f52-790b28d9bb74-kube-api-access-gxtgk\") pod \"auto-csr-approver-29555670-v8lqv\" (UID: \"daa768ee-1644-40fa-8f52-790b28d9bb74\") " pod="openshift-infra/auto-csr-approver-29555670-v8lqv" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.191868 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.192456 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.195348 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555670-v8lqv"] Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.206288 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx"] Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.289491 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxtgk\" (UniqueName: \"kubernetes.io/projected/daa768ee-1644-40fa-8f52-790b28d9bb74-kube-api-access-gxtgk\") pod \"auto-csr-approver-29555670-v8lqv\" (UID: \"daa768ee-1644-40fa-8f52-790b28d9bb74\") " pod="openshift-infra/auto-csr-approver-29555670-v8lqv" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.289590 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a960be2-da37-4f11-a8ba-d2c056550139-config-volume\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.289717 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a960be2-da37-4f11-a8ba-d2c056550139-secret-volume\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.289792 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwtbg\" (UniqueName: \"kubernetes.io/projected/6a960be2-da37-4f11-a8ba-d2c056550139-kube-api-access-jwtbg\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.312510 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxtgk\" (UniqueName: \"kubernetes.io/projected/daa768ee-1644-40fa-8f52-790b28d9bb74-kube-api-access-gxtgk\") pod \"auto-csr-approver-29555670-v8lqv\" (UID: \"daa768ee-1644-40fa-8f52-790b28d9bb74\") " pod="openshift-infra/auto-csr-approver-29555670-v8lqv" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.361233 4926 generic.go:334] "Generic (PLEG): container finished" podID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerID="a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11" exitCode=0 Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.361313 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29mgs" event={"ID":"86dc3218-ffd0-4d12-af08-2c1323f6b670","Type":"ContainerDied","Data":"a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11"} Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.361361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29mgs" event={"ID":"86dc3218-ffd0-4d12-af08-2c1323f6b670","Type":"ContainerStarted","Data":"7e4eb6cfcd4ff4c95aa64507913fece03b2186570bccad32ac581e160aa57575"} Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.394300 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a960be2-da37-4f11-a8ba-d2c056550139-secret-volume\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.395365 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwtbg\" (UniqueName: \"kubernetes.io/projected/6a960be2-da37-4f11-a8ba-d2c056550139-kube-api-access-jwtbg\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.395547 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a960be2-da37-4f11-a8ba-d2c056550139-config-volume\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.396937 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a960be2-da37-4f11-a8ba-d2c056550139-config-volume\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.400107 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a960be2-da37-4f11-a8ba-d2c056550139-secret-volume\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.416178 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwtbg\" (UniqueName: \"kubernetes.io/projected/6a960be2-da37-4f11-a8ba-d2c056550139-kube-api-access-jwtbg\") pod \"collect-profiles-29555670-fm7kx\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.503786 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555670-v8lqv" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.517597 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:00 crc kubenswrapper[4926]: I0312 18:30:00.994686 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555670-v8lqv"] Mar 12 18:30:01 crc kubenswrapper[4926]: W0312 18:30:01.001057 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaa768ee_1644_40fa_8f52_790b28d9bb74.slice/crio-1249afae3d68fc8aea5725888627c7574591fd0ccbc7538d07300e420c6955ba WatchSource:0}: Error finding container 1249afae3d68fc8aea5725888627c7574591fd0ccbc7538d07300e420c6955ba: Status 404 returned error can't find the container with id 1249afae3d68fc8aea5725888627c7574591fd0ccbc7538d07300e420c6955ba Mar 12 18:30:01 crc kubenswrapper[4926]: W0312 18:30:01.089983 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a960be2_da37_4f11_a8ba_d2c056550139.slice/crio-a6a819de11d45bbd132f324f357d9013738950f71532f8e7b0e6fea145687b51 WatchSource:0}: Error finding container a6a819de11d45bbd132f324f357d9013738950f71532f8e7b0e6fea145687b51: Status 404 returned error can't find the container with id a6a819de11d45bbd132f324f357d9013738950f71532f8e7b0e6fea145687b51 Mar 12 18:30:01 crc kubenswrapper[4926]: I0312 18:30:01.092307 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx"] Mar 12 18:30:01 crc kubenswrapper[4926]: I0312 18:30:01.375601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" event={"ID":"6a960be2-da37-4f11-a8ba-d2c056550139","Type":"ContainerStarted","Data":"fc548ee0da579e49aeea484208cb8bc3f74184e26c19a3b70e464f8de8ed58e9"} Mar 12 18:30:01 crc kubenswrapper[4926]: I0312 18:30:01.375931 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" event={"ID":"6a960be2-da37-4f11-a8ba-d2c056550139","Type":"ContainerStarted","Data":"a6a819de11d45bbd132f324f357d9013738950f71532f8e7b0e6fea145687b51"} Mar 12 18:30:01 crc kubenswrapper[4926]: I0312 18:30:01.377698 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555670-v8lqv" event={"ID":"daa768ee-1644-40fa-8f52-790b28d9bb74","Type":"ContainerStarted","Data":"1249afae3d68fc8aea5725888627c7574591fd0ccbc7538d07300e420c6955ba"} Mar 12 18:30:01 crc kubenswrapper[4926]: I0312 18:30:01.383387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29mgs" event={"ID":"86dc3218-ffd0-4d12-af08-2c1323f6b670","Type":"ContainerStarted","Data":"cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345"} Mar 12 18:30:01 crc kubenswrapper[4926]: I0312 18:30:01.396280 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" podStartSLOduration=1.396258495 podStartE2EDuration="1.396258495s" podCreationTimestamp="2026-03-12 18:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 18:30:01.391609251 +0000 UTC m=+1641.760235584" watchObservedRunningTime="2026-03-12 18:30:01.396258495 +0000 UTC m=+1641.764884828" Mar 12 18:30:02 crc kubenswrapper[4926]: I0312 18:30:02.400782 4926 generic.go:334] "Generic (PLEG): container finished" podID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerID="cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345" exitCode=0 Mar 12 18:30:02 crc kubenswrapper[4926]: I0312 18:30:02.400902 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29mgs" event={"ID":"86dc3218-ffd0-4d12-af08-2c1323f6b670","Type":"ContainerDied","Data":"cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345"} Mar 12 18:30:02 crc kubenswrapper[4926]: I0312 18:30:02.409159 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a960be2-da37-4f11-a8ba-d2c056550139" containerID="fc548ee0da579e49aeea484208cb8bc3f74184e26c19a3b70e464f8de8ed58e9" exitCode=0 Mar 12 18:30:02 crc kubenswrapper[4926]: I0312 18:30:02.409233 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" event={"ID":"6a960be2-da37-4f11-a8ba-d2c056550139","Type":"ContainerDied","Data":"fc548ee0da579e49aeea484208cb8bc3f74184e26c19a3b70e464f8de8ed58e9"} Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.434522 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29mgs" event={"ID":"86dc3218-ffd0-4d12-af08-2c1323f6b670","Type":"ContainerStarted","Data":"8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33"} Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.465459 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-29mgs" podStartSLOduration=1.946480881 podStartE2EDuration="4.465426322s" podCreationTimestamp="2026-03-12 18:29:59 +0000 UTC" firstStartedPulling="2026-03-12 18:30:00.364167744 +0000 UTC m=+1640.732794107" lastFinishedPulling="2026-03-12 18:30:02.883113215 +0000 UTC m=+1643.251739548" observedRunningTime="2026-03-12 18:30:03.464494414 +0000 UTC m=+1643.833120797" watchObservedRunningTime="2026-03-12 18:30:03.465426322 +0000 UTC m=+1643.834052645" Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.830786 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.977863 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a960be2-da37-4f11-a8ba-d2c056550139-secret-volume\") pod \"6a960be2-da37-4f11-a8ba-d2c056550139\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.977932 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a960be2-da37-4f11-a8ba-d2c056550139-config-volume\") pod \"6a960be2-da37-4f11-a8ba-d2c056550139\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.977963 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwtbg\" (UniqueName: \"kubernetes.io/projected/6a960be2-da37-4f11-a8ba-d2c056550139-kube-api-access-jwtbg\") pod \"6a960be2-da37-4f11-a8ba-d2c056550139\" (UID: \"6a960be2-da37-4f11-a8ba-d2c056550139\") " Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.980594 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a960be2-da37-4f11-a8ba-d2c056550139-config-volume" (OuterVolumeSpecName: "config-volume") pod "6a960be2-da37-4f11-a8ba-d2c056550139" (UID: "6a960be2-da37-4f11-a8ba-d2c056550139"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 18:30:03 crc kubenswrapper[4926]: I0312 18:30:03.989665 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a960be2-da37-4f11-a8ba-d2c056550139-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6a960be2-da37-4f11-a8ba-d2c056550139" (UID: "6a960be2-da37-4f11-a8ba-d2c056550139"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.022715 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a960be2-da37-4f11-a8ba-d2c056550139-kube-api-access-jwtbg" (OuterVolumeSpecName: "kube-api-access-jwtbg") pod "6a960be2-da37-4f11-a8ba-d2c056550139" (UID: "6a960be2-da37-4f11-a8ba-d2c056550139"). InnerVolumeSpecName "kube-api-access-jwtbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.080628 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6a960be2-da37-4f11-a8ba-d2c056550139-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.080688 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6a960be2-da37-4f11-a8ba-d2c056550139-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.080701 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwtbg\" (UniqueName: \"kubernetes.io/projected/6a960be2-da37-4f11-a8ba-d2c056550139-kube-api-access-jwtbg\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.443145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" event={"ID":"6a960be2-da37-4f11-a8ba-d2c056550139","Type":"ContainerDied","Data":"a6a819de11d45bbd132f324f357d9013738950f71532f8e7b0e6fea145687b51"} Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.443557 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6a819de11d45bbd132f324f357d9013738950f71532f8e7b0e6fea145687b51" Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.443205 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29555670-fm7kx" Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.445729 4926 generic.go:334] "Generic (PLEG): container finished" podID="daa768ee-1644-40fa-8f52-790b28d9bb74" containerID="799ca781836933eed5f381ea7d6b69d2905b6867e2a978eb9f38bd2b549ed6cb" exitCode=0 Mar 12 18:30:04 crc kubenswrapper[4926]: I0312 18:30:04.446603 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555670-v8lqv" event={"ID":"daa768ee-1644-40fa-8f52-790b28d9bb74","Type":"ContainerDied","Data":"799ca781836933eed5f381ea7d6b69d2905b6867e2a978eb9f38bd2b549ed6cb"} Mar 12 18:30:05 crc kubenswrapper[4926]: I0312 18:30:05.833139 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555670-v8lqv" Mar 12 18:30:05 crc kubenswrapper[4926]: I0312 18:30:05.918385 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxtgk\" (UniqueName: \"kubernetes.io/projected/daa768ee-1644-40fa-8f52-790b28d9bb74-kube-api-access-gxtgk\") pod \"daa768ee-1644-40fa-8f52-790b28d9bb74\" (UID: \"daa768ee-1644-40fa-8f52-790b28d9bb74\") " Mar 12 18:30:05 crc kubenswrapper[4926]: I0312 18:30:05.926270 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daa768ee-1644-40fa-8f52-790b28d9bb74-kube-api-access-gxtgk" (OuterVolumeSpecName: "kube-api-access-gxtgk") pod "daa768ee-1644-40fa-8f52-790b28d9bb74" (UID: "daa768ee-1644-40fa-8f52-790b28d9bb74"). InnerVolumeSpecName "kube-api-access-gxtgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:30:06 crc kubenswrapper[4926]: I0312 18:30:06.021140 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxtgk\" (UniqueName: \"kubernetes.io/projected/daa768ee-1644-40fa-8f52-790b28d9bb74-kube-api-access-gxtgk\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:06 crc kubenswrapper[4926]: I0312 18:30:06.466628 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555670-v8lqv" event={"ID":"daa768ee-1644-40fa-8f52-790b28d9bb74","Type":"ContainerDied","Data":"1249afae3d68fc8aea5725888627c7574591fd0ccbc7538d07300e420c6955ba"} Mar 12 18:30:06 crc kubenswrapper[4926]: I0312 18:30:06.466941 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1249afae3d68fc8aea5725888627c7574591fd0ccbc7538d07300e420c6955ba" Mar 12 18:30:06 crc kubenswrapper[4926]: I0312 18:30:06.466770 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555670-v8lqv" Mar 12 18:30:06 crc kubenswrapper[4926]: I0312 18:30:06.913213 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555664-xqlj7"] Mar 12 18:30:06 crc kubenswrapper[4926]: I0312 18:30:06.923820 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555664-xqlj7"] Mar 12 18:30:08 crc kubenswrapper[4926]: I0312 18:30:08.504217 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0a93b50-2038-4bf8-8c5f-bc77148d55f8" path="/var/lib/kubelet/pods/d0a93b50-2038-4bf8-8c5f-bc77148d55f8/volumes" Mar 12 18:30:09 crc kubenswrapper[4926]: I0312 18:30:09.392944 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:30:09 crc kubenswrapper[4926]: I0312 18:30:09.393114 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:30:09 crc kubenswrapper[4926]: I0312 18:30:09.454302 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:30:09 crc kubenswrapper[4926]: I0312 18:30:09.567211 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:30:09 crc kubenswrapper[4926]: I0312 18:30:09.688830 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29mgs"] Mar 12 18:30:11 crc kubenswrapper[4926]: I0312 18:30:11.514068 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-29mgs" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="registry-server" containerID="cri-o://8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33" gracePeriod=2 Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.527143 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.528187 4926 generic.go:334] "Generic (PLEG): container finished" podID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerID="8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33" exitCode=0 Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.528231 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29mgs" event={"ID":"86dc3218-ffd0-4d12-af08-2c1323f6b670","Type":"ContainerDied","Data":"8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33"} Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.528258 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29mgs" event={"ID":"86dc3218-ffd0-4d12-af08-2c1323f6b670","Type":"ContainerDied","Data":"7e4eb6cfcd4ff4c95aa64507913fece03b2186570bccad32ac581e160aa57575"} Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.528275 4926 scope.go:117] "RemoveContainer" containerID="8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.548746 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-catalog-content\") pod \"86dc3218-ffd0-4d12-af08-2c1323f6b670\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.548824 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-utilities\") pod \"86dc3218-ffd0-4d12-af08-2c1323f6b670\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.548983 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g5dw\" (UniqueName: \"kubernetes.io/projected/86dc3218-ffd0-4d12-af08-2c1323f6b670-kube-api-access-5g5dw\") pod \"86dc3218-ffd0-4d12-af08-2c1323f6b670\" (UID: \"86dc3218-ffd0-4d12-af08-2c1323f6b670\") " Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.551623 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-utilities" (OuterVolumeSpecName: "utilities") pod "86dc3218-ffd0-4d12-af08-2c1323f6b670" (UID: "86dc3218-ffd0-4d12-af08-2c1323f6b670"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.557799 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86dc3218-ffd0-4d12-af08-2c1323f6b670-kube-api-access-5g5dw" (OuterVolumeSpecName: "kube-api-access-5g5dw") pod "86dc3218-ffd0-4d12-af08-2c1323f6b670" (UID: "86dc3218-ffd0-4d12-af08-2c1323f6b670"). InnerVolumeSpecName "kube-api-access-5g5dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.564272 4926 scope.go:117] "RemoveContainer" containerID="cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.576852 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86dc3218-ffd0-4d12-af08-2c1323f6b670" (UID: "86dc3218-ffd0-4d12-af08-2c1323f6b670"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.637398 4926 scope.go:117] "RemoveContainer" containerID="a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.654259 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g5dw\" (UniqueName: \"kubernetes.io/projected/86dc3218-ffd0-4d12-af08-2c1323f6b670-kube-api-access-5g5dw\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.654303 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.654318 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86dc3218-ffd0-4d12-af08-2c1323f6b670-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.677397 4926 scope.go:117] "RemoveContainer" containerID="8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33" Mar 12 18:30:12 crc kubenswrapper[4926]: E0312 18:30:12.678803 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33\": container with ID starting with 8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33 not found: ID does not exist" containerID="8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.678836 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33"} err="failed to get container status \"8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33\": rpc error: code = NotFound desc = could not find container \"8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33\": container with ID starting with 8838f3b187da1dc89cc4a1ec3bc43e78dc94dbbd93ce731af57b9f4fa5832c33 not found: ID does not exist" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.678855 4926 scope.go:117] "RemoveContainer" containerID="cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345" Mar 12 18:30:12 crc kubenswrapper[4926]: E0312 18:30:12.679092 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345\": container with ID starting with cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345 not found: ID does not exist" containerID="cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.679119 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345"} err="failed to get container status \"cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345\": rpc error: code = NotFound desc = could not find container \"cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345\": container with ID starting with cb5751d1b0b2b4518890d4be951952deff5a2671bf075c4b9c9dd65a74f33345 not found: ID does not exist" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.679137 4926 scope.go:117] "RemoveContainer" containerID="a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11" Mar 12 18:30:12 crc kubenswrapper[4926]: E0312 18:30:12.679394 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11\": container with ID starting with a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11 not found: ID does not exist" containerID="a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11" Mar 12 18:30:12 crc kubenswrapper[4926]: I0312 18:30:12.679428 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11"} err="failed to get container status \"a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11\": rpc error: code = NotFound desc = could not find container \"a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11\": container with ID starting with a0856534fb16cefe399c293a5eafbda265f3f89c4fdfbdb5141264cad4c82e11 not found: ID does not exist" Mar 12 18:30:13 crc kubenswrapper[4926]: I0312 18:30:13.540534 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29mgs" Mar 12 18:30:13 crc kubenswrapper[4926]: I0312 18:30:13.586367 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29mgs"] Mar 12 18:30:13 crc kubenswrapper[4926]: I0312 18:30:13.598278 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-29mgs"] Mar 12 18:30:14 crc kubenswrapper[4926]: I0312 18:30:14.505028 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" path="/var/lib/kubelet/pods/86dc3218-ffd0-4d12-af08-2c1323f6b670/volumes" Mar 12 18:30:15 crc kubenswrapper[4926]: I0312 18:30:15.058361 4926 scope.go:117] "RemoveContainer" containerID="e0b6e5ac15bb107790c6fb8dd27667c6693088633c203e8fcc884a1e78a79525" Mar 12 18:30:15 crc kubenswrapper[4926]: I0312 18:30:15.121899 4926 scope.go:117] "RemoveContainer" containerID="3e365f4cb50835b1180ff7c19d7fc397df5fff675446efdb00f5237496ddec2b" Mar 12 18:30:15 crc kubenswrapper[4926]: I0312 18:30:15.150874 4926 scope.go:117] "RemoveContainer" containerID="48ea775ffdf8bca7487b51cdb7c8d987f732c9702e1c95ad138a8eebbbab7c90" Mar 12 18:30:15 crc kubenswrapper[4926]: I0312 18:30:15.179388 4926 scope.go:117] "RemoveContainer" containerID="293a2572fcb8b2d40b187537e1c594705c232add90d5f330faf45d6e18e49dd0" Mar 12 18:30:15 crc kubenswrapper[4926]: I0312 18:30:15.218563 4926 scope.go:117] "RemoveContainer" containerID="a12ed3e178db267afb04649b5718c1d2bcc9a78faa88e5e9519c5b32b47a1362" Mar 12 18:30:30 crc kubenswrapper[4926]: I0312 18:30:30.723506 4926 generic.go:334] "Generic (PLEG): container finished" podID="75a3208b-42f5-412e-a503-ac328f7d9967" containerID="6e0e0f76508b934742dbabb4b94eeab9a4fa625bd527c90d90fa0cda46ddea4a" exitCode=0 Mar 12 18:30:30 crc kubenswrapper[4926]: I0312 18:30:30.723588 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" event={"ID":"75a3208b-42f5-412e-a503-ac328f7d9967","Type":"ContainerDied","Data":"6e0e0f76508b934742dbabb4b94eeab9a4fa625bd527c90d90fa0cda46ddea4a"} Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.172774 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.347728 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/75a3208b-42f5-412e-a503-ac328f7d9967-kube-api-access-mpczd\") pod \"75a3208b-42f5-412e-a503-ac328f7d9967\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.347823 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-ssh-key-openstack-edpm-ipam\") pod \"75a3208b-42f5-412e-a503-ac328f7d9967\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.347897 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-inventory\") pod \"75a3208b-42f5-412e-a503-ac328f7d9967\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.347956 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-bootstrap-combined-ca-bundle\") pod \"75a3208b-42f5-412e-a503-ac328f7d9967\" (UID: \"75a3208b-42f5-412e-a503-ac328f7d9967\") " Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.354030 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "75a3208b-42f5-412e-a503-ac328f7d9967" (UID: "75a3208b-42f5-412e-a503-ac328f7d9967"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.354825 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75a3208b-42f5-412e-a503-ac328f7d9967-kube-api-access-mpczd" (OuterVolumeSpecName: "kube-api-access-mpczd") pod "75a3208b-42f5-412e-a503-ac328f7d9967" (UID: "75a3208b-42f5-412e-a503-ac328f7d9967"). InnerVolumeSpecName "kube-api-access-mpczd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.381405 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-inventory" (OuterVolumeSpecName: "inventory") pod "75a3208b-42f5-412e-a503-ac328f7d9967" (UID: "75a3208b-42f5-412e-a503-ac328f7d9967"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.400896 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "75a3208b-42f5-412e-a503-ac328f7d9967" (UID: "75a3208b-42f5-412e-a503-ac328f7d9967"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.449774 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.449808 4926 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.449818 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpczd\" (UniqueName: \"kubernetes.io/projected/75a3208b-42f5-412e-a503-ac328f7d9967-kube-api-access-mpczd\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.449827 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75a3208b-42f5-412e-a503-ac328f7d9967-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.748211 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" event={"ID":"75a3208b-42f5-412e-a503-ac328f7d9967","Type":"ContainerDied","Data":"beacf7f2c4ecf74d1f45669096c28cb46b36b877d88778fb003f1ce1efb01ae4"} Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.748258 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="beacf7f2c4ecf74d1f45669096c28cb46b36b877d88778fb003f1ce1efb01ae4" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.748407 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.838930 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq"] Mar 12 18:30:32 crc kubenswrapper[4926]: E0312 18:30:32.839511 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="extract-utilities" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839538 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="extract-utilities" Mar 12 18:30:32 crc kubenswrapper[4926]: E0312 18:30:32.839557 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="registry-server" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839567 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="registry-server" Mar 12 18:30:32 crc kubenswrapper[4926]: E0312 18:30:32.839608 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="extract-content" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839618 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="extract-content" Mar 12 18:30:32 crc kubenswrapper[4926]: E0312 18:30:32.839637 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a3208b-42f5-412e-a503-ac328f7d9967" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839646 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a3208b-42f5-412e-a503-ac328f7d9967" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 12 18:30:32 crc kubenswrapper[4926]: E0312 18:30:32.839666 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daa768ee-1644-40fa-8f52-790b28d9bb74" containerName="oc" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839675 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="daa768ee-1644-40fa-8f52-790b28d9bb74" containerName="oc" Mar 12 18:30:32 crc kubenswrapper[4926]: E0312 18:30:32.839693 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a960be2-da37-4f11-a8ba-d2c056550139" containerName="collect-profiles" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839701 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a960be2-da37-4f11-a8ba-d2c056550139" containerName="collect-profiles" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839930 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a960be2-da37-4f11-a8ba-d2c056550139" containerName="collect-profiles" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839956 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="75a3208b-42f5-412e-a503-ac328f7d9967" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839977 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="daa768ee-1644-40fa-8f52-790b28d9bb74" containerName="oc" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.839998 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="86dc3218-ffd0-4d12-af08-2c1323f6b670" containerName="registry-server" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.840857 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.843577 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.843719 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.843738 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.846902 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq"] Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.849237 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.963373 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5q9m\" (UniqueName: \"kubernetes.io/projected/eab3d0c2-5edc-4657-928f-52a87de2293a-kube-api-access-k5q9m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.963636 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:32 crc kubenswrapper[4926]: I0312 18:30:32.963840 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:33 crc kubenswrapper[4926]: I0312 18:30:33.065314 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5q9m\" (UniqueName: \"kubernetes.io/projected/eab3d0c2-5edc-4657-928f-52a87de2293a-kube-api-access-k5q9m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:33 crc kubenswrapper[4926]: I0312 18:30:33.065540 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:33 crc kubenswrapper[4926]: I0312 18:30:33.065638 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:33 crc kubenswrapper[4926]: I0312 18:30:33.071288 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:33 crc kubenswrapper[4926]: I0312 18:30:33.078914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:33 crc kubenswrapper[4926]: I0312 18:30:33.095252 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5q9m\" (UniqueName: \"kubernetes.io/projected/eab3d0c2-5edc-4657-928f-52a87de2293a-kube-api-access-k5q9m\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:33 crc kubenswrapper[4926]: I0312 18:30:33.166095 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:30:34 crc kubenswrapper[4926]: I0312 18:30:34.321102 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq"] Mar 12 18:30:34 crc kubenswrapper[4926]: I0312 18:30:34.769220 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" event={"ID":"eab3d0c2-5edc-4657-928f-52a87de2293a","Type":"ContainerStarted","Data":"b5e1946b94fd7ed87879292f5fef7a875c3edb29871fbcc32b4cd0930bf71f7f"} Mar 12 18:30:35 crc kubenswrapper[4926]: I0312 18:30:35.782931 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" event={"ID":"eab3d0c2-5edc-4657-928f-52a87de2293a","Type":"ContainerStarted","Data":"d773dc87aff248613f6b9d5825635f53e85aba8c66114657cd4d6a99eafa2f07"} Mar 12 18:30:35 crc kubenswrapper[4926]: I0312 18:30:35.813951 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" podStartSLOduration=3.149726207 podStartE2EDuration="3.813926245s" podCreationTimestamp="2026-03-12 18:30:32 +0000 UTC" firstStartedPulling="2026-03-12 18:30:34.323212945 +0000 UTC m=+1674.691839278" lastFinishedPulling="2026-03-12 18:30:34.987412973 +0000 UTC m=+1675.356039316" observedRunningTime="2026-03-12 18:30:35.8053592 +0000 UTC m=+1676.173985543" watchObservedRunningTime="2026-03-12 18:30:35.813926245 +0000 UTC m=+1676.182552588" Mar 12 18:30:56 crc kubenswrapper[4926]: I0312 18:30:56.818260 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:30:56 crc kubenswrapper[4926]: I0312 18:30:56.819087 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:31:15 crc kubenswrapper[4926]: I0312 18:31:15.428734 4926 scope.go:117] "RemoveContainer" containerID="7655ace271d6f3feff7f2f4eaa16c5ea625e5fe6beab0c4038907b7f5e715987" Mar 12 18:31:15 crc kubenswrapper[4926]: I0312 18:31:15.455164 4926 scope.go:117] "RemoveContainer" containerID="435ef2132df5554295889ec9ba808434afc58d13077a9a39e7222dd605149d55" Mar 12 18:31:15 crc kubenswrapper[4926]: I0312 18:31:15.479871 4926 scope.go:117] "RemoveContainer" containerID="89db61a8e375fc043c0de0eeea971f98c1fb8c205fa68c44c6f73887667808c6" Mar 12 18:31:15 crc kubenswrapper[4926]: I0312 18:31:15.501599 4926 scope.go:117] "RemoveContainer" containerID="d312e71f700c8194066a4bfb0efbe92c4e2b9fbced6805317853fff15c35d5d6" Mar 12 18:31:15 crc kubenswrapper[4926]: I0312 18:31:15.524294 4926 scope.go:117] "RemoveContainer" containerID="0b78f1753e8c4a22ed2f4147f06789b56f6db668b7a92626590701ea109bcd09" Mar 12 18:31:15 crc kubenswrapper[4926]: I0312 18:31:15.542320 4926 scope.go:117] "RemoveContainer" containerID="2acf5477ea08c1fb97c59765c73d7abfdadc6189fbd1cf4549d66c1007c1d817" Mar 12 18:31:26 crc kubenswrapper[4926]: I0312 18:31:26.817654 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:31:26 crc kubenswrapper[4926]: I0312 18:31:26.818350 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:31:42 crc kubenswrapper[4926]: I0312 18:31:42.545860 4926 generic.go:334] "Generic (PLEG): container finished" podID="eab3d0c2-5edc-4657-928f-52a87de2293a" containerID="d773dc87aff248613f6b9d5825635f53e85aba8c66114657cd4d6a99eafa2f07" exitCode=0 Mar 12 18:31:42 crc kubenswrapper[4926]: I0312 18:31:42.545979 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" event={"ID":"eab3d0c2-5edc-4657-928f-52a87de2293a","Type":"ContainerDied","Data":"d773dc87aff248613f6b9d5825635f53e85aba8c66114657cd4d6a99eafa2f07"} Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.060671 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.214557 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5q9m\" (UniqueName: \"kubernetes.io/projected/eab3d0c2-5edc-4657-928f-52a87de2293a-kube-api-access-k5q9m\") pod \"eab3d0c2-5edc-4657-928f-52a87de2293a\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.214617 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-ssh-key-openstack-edpm-ipam\") pod \"eab3d0c2-5edc-4657-928f-52a87de2293a\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.214798 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-inventory\") pod \"eab3d0c2-5edc-4657-928f-52a87de2293a\" (UID: \"eab3d0c2-5edc-4657-928f-52a87de2293a\") " Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.222152 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab3d0c2-5edc-4657-928f-52a87de2293a-kube-api-access-k5q9m" (OuterVolumeSpecName: "kube-api-access-k5q9m") pod "eab3d0c2-5edc-4657-928f-52a87de2293a" (UID: "eab3d0c2-5edc-4657-928f-52a87de2293a"). InnerVolumeSpecName "kube-api-access-k5q9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.243854 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-inventory" (OuterVolumeSpecName: "inventory") pod "eab3d0c2-5edc-4657-928f-52a87de2293a" (UID: "eab3d0c2-5edc-4657-928f-52a87de2293a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.270304 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "eab3d0c2-5edc-4657-928f-52a87de2293a" (UID: "eab3d0c2-5edc-4657-928f-52a87de2293a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.317423 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5q9m\" (UniqueName: \"kubernetes.io/projected/eab3d0c2-5edc-4657-928f-52a87de2293a-kube-api-access-k5q9m\") on node \"crc\" DevicePath \"\"" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.317471 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.317486 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eab3d0c2-5edc-4657-928f-52a87de2293a-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.573965 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" event={"ID":"eab3d0c2-5edc-4657-928f-52a87de2293a","Type":"ContainerDied","Data":"b5e1946b94fd7ed87879292f5fef7a875c3edb29871fbcc32b4cd0930bf71f7f"} Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.574471 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5e1946b94fd7ed87879292f5fef7a875c3edb29871fbcc32b4cd0930bf71f7f" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.574076 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.733082 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj"] Mar 12 18:31:44 crc kubenswrapper[4926]: E0312 18:31:44.733526 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab3d0c2-5edc-4657-928f-52a87de2293a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.733545 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab3d0c2-5edc-4657-928f-52a87de2293a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.733745 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab3d0c2-5edc-4657-928f-52a87de2293a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.734293 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.736324 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.736353 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.736756 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.741959 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.751797 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj"] Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.827250 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd4mv\" (UniqueName: \"kubernetes.io/projected/96a114de-73e3-4088-902e-c3d1fcaaa3ad-kube-api-access-wd4mv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.827295 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.827325 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.929294 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd4mv\" (UniqueName: \"kubernetes.io/projected/96a114de-73e3-4088-902e-c3d1fcaaa3ad-kube-api-access-wd4mv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.929373 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.929497 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.936569 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.937639 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:44 crc kubenswrapper[4926]: I0312 18:31:44.951142 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd4mv\" (UniqueName: \"kubernetes.io/projected/96a114de-73e3-4088-902e-c3d1fcaaa3ad-kube-api-access-wd4mv\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-55zdj\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:45 crc kubenswrapper[4926]: I0312 18:31:45.054640 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:45 crc kubenswrapper[4926]: I0312 18:31:45.651016 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj"] Mar 12 18:31:46 crc kubenswrapper[4926]: I0312 18:31:46.601375 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" event={"ID":"96a114de-73e3-4088-902e-c3d1fcaaa3ad","Type":"ContainerStarted","Data":"710f1a12525865b4371846d75a84287b63e3823a526dd6bb16dfbf5e78d6ddda"} Mar 12 18:31:46 crc kubenswrapper[4926]: I0312 18:31:46.601853 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" event={"ID":"96a114de-73e3-4088-902e-c3d1fcaaa3ad","Type":"ContainerStarted","Data":"e75f8ec3b61a16b23223b1074cf55568f6ef14af0ee49e12c4225a40410f4cd0"} Mar 12 18:31:46 crc kubenswrapper[4926]: I0312 18:31:46.634359 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" podStartSLOduration=2.134787726 podStartE2EDuration="2.634340266s" podCreationTimestamp="2026-03-12 18:31:44 +0000 UTC" firstStartedPulling="2026-03-12 18:31:45.666642832 +0000 UTC m=+1746.035269175" lastFinishedPulling="2026-03-12 18:31:46.166195372 +0000 UTC m=+1746.534821715" observedRunningTime="2026-03-12 18:31:46.621848648 +0000 UTC m=+1746.990475001" watchObservedRunningTime="2026-03-12 18:31:46.634340266 +0000 UTC m=+1747.002966619" Mar 12 18:31:51 crc kubenswrapper[4926]: I0312 18:31:51.661208 4926 generic.go:334] "Generic (PLEG): container finished" podID="96a114de-73e3-4088-902e-c3d1fcaaa3ad" containerID="710f1a12525865b4371846d75a84287b63e3823a526dd6bb16dfbf5e78d6ddda" exitCode=0 Mar 12 18:31:51 crc kubenswrapper[4926]: I0312 18:31:51.661318 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" event={"ID":"96a114de-73e3-4088-902e-c3d1fcaaa3ad","Type":"ContainerDied","Data":"710f1a12525865b4371846d75a84287b63e3823a526dd6bb16dfbf5e78d6ddda"} Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.115222 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.205311 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-inventory\") pod \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.205508 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-ssh-key-openstack-edpm-ipam\") pod \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.205650 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd4mv\" (UniqueName: \"kubernetes.io/projected/96a114de-73e3-4088-902e-c3d1fcaaa3ad-kube-api-access-wd4mv\") pod \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\" (UID: \"96a114de-73e3-4088-902e-c3d1fcaaa3ad\") " Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.212571 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a114de-73e3-4088-902e-c3d1fcaaa3ad-kube-api-access-wd4mv" (OuterVolumeSpecName: "kube-api-access-wd4mv") pod "96a114de-73e3-4088-902e-c3d1fcaaa3ad" (UID: "96a114de-73e3-4088-902e-c3d1fcaaa3ad"). InnerVolumeSpecName "kube-api-access-wd4mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.235771 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "96a114de-73e3-4088-902e-c3d1fcaaa3ad" (UID: "96a114de-73e3-4088-902e-c3d1fcaaa3ad"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.244280 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-inventory" (OuterVolumeSpecName: "inventory") pod "96a114de-73e3-4088-902e-c3d1fcaaa3ad" (UID: "96a114de-73e3-4088-902e-c3d1fcaaa3ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.308837 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.308881 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96a114de-73e3-4088-902e-c3d1fcaaa3ad-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.308894 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd4mv\" (UniqueName: \"kubernetes.io/projected/96a114de-73e3-4088-902e-c3d1fcaaa3ad-kube-api-access-wd4mv\") on node \"crc\" DevicePath \"\"" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.686057 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" event={"ID":"96a114de-73e3-4088-902e-c3d1fcaaa3ad","Type":"ContainerDied","Data":"e75f8ec3b61a16b23223b1074cf55568f6ef14af0ee49e12c4225a40410f4cd0"} Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.686356 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e75f8ec3b61a16b23223b1074cf55568f6ef14af0ee49e12c4225a40410f4cd0" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.686114 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-55zdj" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.766182 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr"] Mar 12 18:31:53 crc kubenswrapper[4926]: E0312 18:31:53.766843 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a114de-73e3-4088-902e-c3d1fcaaa3ad" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.766863 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a114de-73e3-4088-902e-c3d1fcaaa3ad" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.767047 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a114de-73e3-4088-902e-c3d1fcaaa3ad" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.767666 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.770914 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.771073 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.771344 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.773195 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.808247 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr"] Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.920578 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.920652 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:53 crc kubenswrapper[4926]: I0312 18:31:53.921253 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r9cp\" (UniqueName: \"kubernetes.io/projected/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-kube-api-access-7r9cp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.023145 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r9cp\" (UniqueName: \"kubernetes.io/projected/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-kube-api-access-7r9cp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.023350 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.023391 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.028279 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.031279 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.046628 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r9cp\" (UniqueName: \"kubernetes.io/projected/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-kube-api-access-7r9cp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-njwrr\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.122658 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:31:54 crc kubenswrapper[4926]: I0312 18:31:54.759256 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr"] Mar 12 18:31:55 crc kubenswrapper[4926]: I0312 18:31:55.715354 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" event={"ID":"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46","Type":"ContainerStarted","Data":"573a9ceb3cb256f42bb14e281728fe4976f6e1621ed33c7cafc9ec1b14a4b12b"} Mar 12 18:31:55 crc kubenswrapper[4926]: I0312 18:31:55.715749 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" event={"ID":"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46","Type":"ContainerStarted","Data":"2a98a7c2f2a3e2f97eddca16c9e2490f09751655edf87b4db97219d94185a80f"} Mar 12 18:31:55 crc kubenswrapper[4926]: I0312 18:31:55.738902 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" podStartSLOduration=2.339037015 podStartE2EDuration="2.73887545s" podCreationTimestamp="2026-03-12 18:31:53 +0000 UTC" firstStartedPulling="2026-03-12 18:31:54.747128702 +0000 UTC m=+1755.115755035" lastFinishedPulling="2026-03-12 18:31:55.146967127 +0000 UTC m=+1755.515593470" observedRunningTime="2026-03-12 18:31:55.733538565 +0000 UTC m=+1756.102164898" watchObservedRunningTime="2026-03-12 18:31:55.73887545 +0000 UTC m=+1756.107501813" Mar 12 18:31:56 crc kubenswrapper[4926]: I0312 18:31:56.817998 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:31:56 crc kubenswrapper[4926]: I0312 18:31:56.818374 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:31:56 crc kubenswrapper[4926]: I0312 18:31:56.818461 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:31:56 crc kubenswrapper[4926]: I0312 18:31:56.819576 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:31:56 crc kubenswrapper[4926]: I0312 18:31:56.819686 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" gracePeriod=600 Mar 12 18:31:56 crc kubenswrapper[4926]: E0312 18:31:56.950871 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:31:57 crc kubenswrapper[4926]: I0312 18:31:57.750119 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" exitCode=0 Mar 12 18:31:57 crc kubenswrapper[4926]: I0312 18:31:57.750189 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e"} Mar 12 18:31:57 crc kubenswrapper[4926]: I0312 18:31:57.750244 4926 scope.go:117] "RemoveContainer" containerID="759fd18072cdf8fcc7bc2d92cc950b5720a437d7e4487f5098fffd2244e21cde" Mar 12 18:31:57 crc kubenswrapper[4926]: I0312 18:31:57.751147 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:31:57 crc kubenswrapper[4926]: E0312 18:31:57.751760 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.131914 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555672-dq68h"] Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.133671 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555672-dq68h" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.136046 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.136204 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.136352 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.148692 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555672-dq68h"] Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.245812 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hth86\" (UniqueName: \"kubernetes.io/projected/609b8dbc-517d-4483-b02a-d7445cd2aa2f-kube-api-access-hth86\") pod \"auto-csr-approver-29555672-dq68h\" (UID: \"609b8dbc-517d-4483-b02a-d7445cd2aa2f\") " pod="openshift-infra/auto-csr-approver-29555672-dq68h" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.347788 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hth86\" (UniqueName: \"kubernetes.io/projected/609b8dbc-517d-4483-b02a-d7445cd2aa2f-kube-api-access-hth86\") pod \"auto-csr-approver-29555672-dq68h\" (UID: \"609b8dbc-517d-4483-b02a-d7445cd2aa2f\") " pod="openshift-infra/auto-csr-approver-29555672-dq68h" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.372831 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hth86\" (UniqueName: \"kubernetes.io/projected/609b8dbc-517d-4483-b02a-d7445cd2aa2f-kube-api-access-hth86\") pod \"auto-csr-approver-29555672-dq68h\" (UID: \"609b8dbc-517d-4483-b02a-d7445cd2aa2f\") " pod="openshift-infra/auto-csr-approver-29555672-dq68h" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.451058 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555672-dq68h" Mar 12 18:32:00 crc kubenswrapper[4926]: I0312 18:32:00.948253 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555672-dq68h"] Mar 12 18:32:01 crc kubenswrapper[4926]: I0312 18:32:01.794677 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555672-dq68h" event={"ID":"609b8dbc-517d-4483-b02a-d7445cd2aa2f","Type":"ContainerStarted","Data":"7182f54670fbf2e0ae70b5529a32090424d21ecca05ea5da6c657dbbb51c12ca"} Mar 12 18:32:02 crc kubenswrapper[4926]: I0312 18:32:02.807520 4926 generic.go:334] "Generic (PLEG): container finished" podID="609b8dbc-517d-4483-b02a-d7445cd2aa2f" containerID="95bf3d49a01a19202e8eb63fbef45841e7cc791e3f3ccb3272c728981fc71d7d" exitCode=0 Mar 12 18:32:02 crc kubenswrapper[4926]: I0312 18:32:02.807628 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555672-dq68h" event={"ID":"609b8dbc-517d-4483-b02a-d7445cd2aa2f","Type":"ContainerDied","Data":"95bf3d49a01a19202e8eb63fbef45841e7cc791e3f3ccb3272c728981fc71d7d"} Mar 12 18:32:04 crc kubenswrapper[4926]: I0312 18:32:04.180788 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555672-dq68h" Mar 12 18:32:04 crc kubenswrapper[4926]: I0312 18:32:04.225676 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hth86\" (UniqueName: \"kubernetes.io/projected/609b8dbc-517d-4483-b02a-d7445cd2aa2f-kube-api-access-hth86\") pod \"609b8dbc-517d-4483-b02a-d7445cd2aa2f\" (UID: \"609b8dbc-517d-4483-b02a-d7445cd2aa2f\") " Mar 12 18:32:04 crc kubenswrapper[4926]: I0312 18:32:04.231546 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/609b8dbc-517d-4483-b02a-d7445cd2aa2f-kube-api-access-hth86" (OuterVolumeSpecName: "kube-api-access-hth86") pod "609b8dbc-517d-4483-b02a-d7445cd2aa2f" (UID: "609b8dbc-517d-4483-b02a-d7445cd2aa2f"). InnerVolumeSpecName "kube-api-access-hth86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:32:04 crc kubenswrapper[4926]: I0312 18:32:04.327723 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hth86\" (UniqueName: \"kubernetes.io/projected/609b8dbc-517d-4483-b02a-d7445cd2aa2f-kube-api-access-hth86\") on node \"crc\" DevicePath \"\"" Mar 12 18:32:04 crc kubenswrapper[4926]: I0312 18:32:04.838642 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555672-dq68h" event={"ID":"609b8dbc-517d-4483-b02a-d7445cd2aa2f","Type":"ContainerDied","Data":"7182f54670fbf2e0ae70b5529a32090424d21ecca05ea5da6c657dbbb51c12ca"} Mar 12 18:32:04 crc kubenswrapper[4926]: I0312 18:32:04.838699 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7182f54670fbf2e0ae70b5529a32090424d21ecca05ea5da6c657dbbb51c12ca" Mar 12 18:32:04 crc kubenswrapper[4926]: I0312 18:32:04.838707 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555672-dq68h" Mar 12 18:32:05 crc kubenswrapper[4926]: I0312 18:32:05.273260 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555666-x5zgg"] Mar 12 18:32:05 crc kubenswrapper[4926]: I0312 18:32:05.288328 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555666-x5zgg"] Mar 12 18:32:06 crc kubenswrapper[4926]: I0312 18:32:06.507145 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ff186e2-5cbe-493a-b911-426e982888cb" path="/var/lib/kubelet/pods/5ff186e2-5cbe-493a-b911-426e982888cb/volumes" Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.044172 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-87jsw"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.053315 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gk626"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.065461 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-87jsw"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.074123 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-kslsb"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.082635 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-c596-account-create-update-kw2ql"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.109143 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-15a0-account-create-update-b4wdv"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.124767 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gk626"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.135654 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-c596-account-create-update-kw2ql"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.146499 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-kslsb"] Mar 12 18:32:07 crc kubenswrapper[4926]: I0312 18:32:07.155712 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-15a0-account-create-update-b4wdv"] Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.032493 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-b886-account-create-update-sdqf6"] Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.040592 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-b886-account-create-update-sdqf6"] Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.508428 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e559975-5aca-457e-8c50-465552595381" path="/var/lib/kubelet/pods/1e559975-5aca-457e-8c50-465552595381/volumes" Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.509216 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496" path="/var/lib/kubelet/pods/2e70e0f6-9596-4d7f-bbc6-e5b6e3f6c496/volumes" Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.509974 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f98d642-e5f2-44de-9259-15b5eed6b80c" path="/var/lib/kubelet/pods/3f98d642-e5f2-44de-9259-15b5eed6b80c/volumes" Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.510631 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51bf320d-e5c9-43c6-baf7-e1f2f9ee3313" path="/var/lib/kubelet/pods/51bf320d-e5c9-43c6-baf7-e1f2f9ee3313/volumes" Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.511838 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce8493fd-3e35-41fb-8daa-febd2238ce1b" path="/var/lib/kubelet/pods/ce8493fd-3e35-41fb-8daa-febd2238ce1b/volumes" Mar 12 18:32:08 crc kubenswrapper[4926]: I0312 18:32:08.512498 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb8a8892-82ee-4502-b76d-ca289485809b" path="/var/lib/kubelet/pods/eb8a8892-82ee-4502-b76d-ca289485809b/volumes" Mar 12 18:32:12 crc kubenswrapper[4926]: I0312 18:32:12.490467 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:32:12 crc kubenswrapper[4926]: E0312 18:32:12.491121 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.609037 4926 scope.go:117] "RemoveContainer" containerID="303e0f37b5ff32adc0c5ce3a0b0f3252de8de58704f569f8ef9bdf6d235b21d1" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.680726 4926 scope.go:117] "RemoveContainer" containerID="ed6d1abff4f7da34f442e943b17a81aeae55a22054d604db8a2f0c813a7bb890" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.711503 4926 scope.go:117] "RemoveContainer" containerID="97dffad6ad390315746d80bb46b2150bb89f6d5359806877446913c809356199" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.757397 4926 scope.go:117] "RemoveContainer" containerID="0477440cc1caee314ece47874cba75dc234908b5f0302ff1737f87d526dec4a4" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.817323 4926 scope.go:117] "RemoveContainer" containerID="2d6e21accb5b365373d4e446af9a0d2f3bb681b5f74d1a1d1d5bbeb21b98d345" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.846766 4926 scope.go:117] "RemoveContainer" containerID="00d08df89cba0281fd578a80d34eeb606178fe179a18e0bb4dca41d34fcaae6c" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.875998 4926 scope.go:117] "RemoveContainer" containerID="7093c3070454c81ffcfc913032818ad12b03f50c370515a66252d6a038735742" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.897572 4926 scope.go:117] "RemoveContainer" containerID="e6bde295c7812497b3848b236f4e958ec3015f8ebeca1a4fdbd6ee1413805cb9" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.941096 4926 scope.go:117] "RemoveContainer" containerID="2ba173648181598e4717e6ee2a92ecb7ee6351d447d5a38a4302f5af077322b5" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.965431 4926 scope.go:117] "RemoveContainer" containerID="22a2d1e8587226f797d924ae278b1c0e7b846e0cc05d9f335cbb31ac04af937c" Mar 12 18:32:15 crc kubenswrapper[4926]: I0312 18:32:15.984288 4926 scope.go:117] "RemoveContainer" containerID="d5aba66a32fd03fd9f7cad091c15aa0c8e52be10f9b3914b0869cd5e59465cde" Mar 12 18:32:25 crc kubenswrapper[4926]: I0312 18:32:25.489736 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:32:25 crc kubenswrapper[4926]: E0312 18:32:25.490503 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:32:31 crc kubenswrapper[4926]: I0312 18:32:31.044676 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-jlt78"] Mar 12 18:32:31 crc kubenswrapper[4926]: I0312 18:32:31.054558 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-jlt78"] Mar 12 18:32:31 crc kubenswrapper[4926]: I0312 18:32:31.125721 4926 generic.go:334] "Generic (PLEG): container finished" podID="7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46" containerID="573a9ceb3cb256f42bb14e281728fe4976f6e1621ed33c7cafc9ec1b14a4b12b" exitCode=0 Mar 12 18:32:31 crc kubenswrapper[4926]: I0312 18:32:31.125786 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" event={"ID":"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46","Type":"ContainerDied","Data":"573a9ceb3cb256f42bb14e281728fe4976f6e1621ed33c7cafc9ec1b14a4b12b"} Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.509804 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="349ca4f5-349b-45ab-98a4-844fa00599d0" path="/var/lib/kubelet/pods/349ca4f5-349b-45ab-98a4-844fa00599d0/volumes" Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.601753 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.666995 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-inventory\") pod \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.668403 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-ssh-key-openstack-edpm-ipam\") pod \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.668533 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r9cp\" (UniqueName: \"kubernetes.io/projected/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-kube-api-access-7r9cp\") pod \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\" (UID: \"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46\") " Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.679850 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-kube-api-access-7r9cp" (OuterVolumeSpecName: "kube-api-access-7r9cp") pod "7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46" (UID: "7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46"). InnerVolumeSpecName "kube-api-access-7r9cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.707554 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46" (UID: "7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.707609 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-inventory" (OuterVolumeSpecName: "inventory") pod "7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46" (UID: "7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.781404 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r9cp\" (UniqueName: \"kubernetes.io/projected/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-kube-api-access-7r9cp\") on node \"crc\" DevicePath \"\"" Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.781740 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:32:32 crc kubenswrapper[4926]: I0312 18:32:32.781775 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.144744 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" event={"ID":"7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46","Type":"ContainerDied","Data":"2a98a7c2f2a3e2f97eddca16c9e2490f09751655edf87b4db97219d94185a80f"} Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.144849 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a98a7c2f2a3e2f97eddca16c9e2490f09751655edf87b4db97219d94185a80f" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.144779 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-njwrr" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.239527 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w"] Mar 12 18:32:33 crc kubenswrapper[4926]: E0312 18:32:33.240172 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="609b8dbc-517d-4483-b02a-d7445cd2aa2f" containerName="oc" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.240201 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="609b8dbc-517d-4483-b02a-d7445cd2aa2f" containerName="oc" Mar 12 18:32:33 crc kubenswrapper[4926]: E0312 18:32:33.240241 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.240255 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.240639 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="609b8dbc-517d-4483-b02a-d7445cd2aa2f" containerName="oc" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.240690 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.241648 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.244927 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.245155 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.245878 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.246289 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.254473 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w"] Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.394703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.394799 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.394938 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxv7d\" (UniqueName: \"kubernetes.io/projected/1891d243-9edd-48aa-88ff-f943dc337e8d-kube-api-access-bxv7d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.496743 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.496892 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.497109 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxv7d\" (UniqueName: \"kubernetes.io/projected/1891d243-9edd-48aa-88ff-f943dc337e8d-kube-api-access-bxv7d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.500686 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.506662 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-ssh-key-openstack-edpm-ipam\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.516373 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxv7d\" (UniqueName: \"kubernetes.io/projected/1891d243-9edd-48aa-88ff-f943dc337e8d-kube-api-access-bxv7d\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:33 crc kubenswrapper[4926]: I0312 18:32:33.564819 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:34 crc kubenswrapper[4926]: I0312 18:32:34.035197 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-40f3-account-create-update-6f5kz"] Mar 12 18:32:34 crc kubenswrapper[4926]: I0312 18:32:34.044728 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-n9jj2"] Mar 12 18:32:34 crc kubenswrapper[4926]: I0312 18:32:34.057051 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-40f3-account-create-update-6f5kz"] Mar 12 18:32:34 crc kubenswrapper[4926]: I0312 18:32:34.069320 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-n9jj2"] Mar 12 18:32:34 crc kubenswrapper[4926]: I0312 18:32:34.180596 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w"] Mar 12 18:32:34 crc kubenswrapper[4926]: I0312 18:32:34.503161 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a1a6049-a924-4e9e-adad-e6ec84732eb9" path="/var/lib/kubelet/pods/6a1a6049-a924-4e9e-adad-e6ec84732eb9/volumes" Mar 12 18:32:34 crc kubenswrapper[4926]: I0312 18:32:34.504211 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa60a9bd-1933-44b8-ac17-be1f72c9da68" path="/var/lib/kubelet/pods/fa60a9bd-1933-44b8-ac17-be1f72c9da68/volumes" Mar 12 18:32:35 crc kubenswrapper[4926]: I0312 18:32:35.166094 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" event={"ID":"1891d243-9edd-48aa-88ff-f943dc337e8d","Type":"ContainerStarted","Data":"7548c9cec626747350a0147b1ed113cbc39ffe794cdd24b52386a11968ee94db"} Mar 12 18:32:35 crc kubenswrapper[4926]: I0312 18:32:35.166636 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" event={"ID":"1891d243-9edd-48aa-88ff-f943dc337e8d","Type":"ContainerStarted","Data":"3b48a2d4c059ac3596463fa9774e80dda754bd59c1f5280cd14ce2ca34cfd373"} Mar 12 18:32:35 crc kubenswrapper[4926]: I0312 18:32:35.207774 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" podStartSLOduration=1.699335222 podStartE2EDuration="2.207755986s" podCreationTimestamp="2026-03-12 18:32:33 +0000 UTC" firstStartedPulling="2026-03-12 18:32:34.193711885 +0000 UTC m=+1794.562338218" lastFinishedPulling="2026-03-12 18:32:34.702132649 +0000 UTC m=+1795.070758982" observedRunningTime="2026-03-12 18:32:35.207106896 +0000 UTC m=+1795.575733229" watchObservedRunningTime="2026-03-12 18:32:35.207755986 +0000 UTC m=+1795.576382319" Mar 12 18:32:36 crc kubenswrapper[4926]: I0312 18:32:36.489954 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:32:36 crc kubenswrapper[4926]: E0312 18:32:36.490676 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:32:37 crc kubenswrapper[4926]: I0312 18:32:37.029780 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-2c76-account-create-update-wvgzd"] Mar 12 18:32:37 crc kubenswrapper[4926]: I0312 18:32:37.042472 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-2c76-account-create-update-wvgzd"] Mar 12 18:32:37 crc kubenswrapper[4926]: I0312 18:32:37.053166 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-n28hs"] Mar 12 18:32:37 crc kubenswrapper[4926]: I0312 18:32:37.063578 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-n28hs"] Mar 12 18:32:37 crc kubenswrapper[4926]: I0312 18:32:37.071585 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-bf50-account-create-update-ckzh6"] Mar 12 18:32:37 crc kubenswrapper[4926]: I0312 18:32:37.078921 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-bf50-account-create-update-ckzh6"] Mar 12 18:32:38 crc kubenswrapper[4926]: I0312 18:32:38.033397 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-qr6zk"] Mar 12 18:32:38 crc kubenswrapper[4926]: I0312 18:32:38.043947 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-qr6zk"] Mar 12 18:32:38 crc kubenswrapper[4926]: I0312 18:32:38.510646 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05959443-d099-4653-9736-7745ba1ce331" path="/var/lib/kubelet/pods/05959443-d099-4653-9736-7745ba1ce331/volumes" Mar 12 18:32:38 crc kubenswrapper[4926]: I0312 18:32:38.512078 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e5bf33c-5cb6-4b29-95f4-fcf47183be58" path="/var/lib/kubelet/pods/9e5bf33c-5cb6-4b29-95f4-fcf47183be58/volumes" Mar 12 18:32:38 crc kubenswrapper[4926]: I0312 18:32:38.513212 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea314c69-7524-43d6-9d4f-9fdb16510952" path="/var/lib/kubelet/pods/ea314c69-7524-43d6-9d4f-9fdb16510952/volumes" Mar 12 18:32:38 crc kubenswrapper[4926]: I0312 18:32:38.514342 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7917312-5922-44df-a838-6b5452e8bb84" path="/var/lib/kubelet/pods/f7917312-5922-44df-a838-6b5452e8bb84/volumes" Mar 12 18:32:39 crc kubenswrapper[4926]: I0312 18:32:39.032841 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-72msf"] Mar 12 18:32:39 crc kubenswrapper[4926]: I0312 18:32:39.045040 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-72msf"] Mar 12 18:32:39 crc kubenswrapper[4926]: I0312 18:32:39.206706 4926 generic.go:334] "Generic (PLEG): container finished" podID="1891d243-9edd-48aa-88ff-f943dc337e8d" containerID="7548c9cec626747350a0147b1ed113cbc39ffe794cdd24b52386a11968ee94db" exitCode=0 Mar 12 18:32:39 crc kubenswrapper[4926]: I0312 18:32:39.206776 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" event={"ID":"1891d243-9edd-48aa-88ff-f943dc337e8d","Type":"ContainerDied","Data":"7548c9cec626747350a0147b1ed113cbc39ffe794cdd24b52386a11968ee94db"} Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.512988 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bad0b5a-a817-45ec-9ebb-7b30d7492ed8" path="/var/lib/kubelet/pods/6bad0b5a-a817-45ec-9ebb-7b30d7492ed8/volumes" Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.760172 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.877576 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-inventory\") pod \"1891d243-9edd-48aa-88ff-f943dc337e8d\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.877673 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-ssh-key-openstack-edpm-ipam\") pod \"1891d243-9edd-48aa-88ff-f943dc337e8d\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.877801 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxv7d\" (UniqueName: \"kubernetes.io/projected/1891d243-9edd-48aa-88ff-f943dc337e8d-kube-api-access-bxv7d\") pod \"1891d243-9edd-48aa-88ff-f943dc337e8d\" (UID: \"1891d243-9edd-48aa-88ff-f943dc337e8d\") " Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.885868 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1891d243-9edd-48aa-88ff-f943dc337e8d-kube-api-access-bxv7d" (OuterVolumeSpecName: "kube-api-access-bxv7d") pod "1891d243-9edd-48aa-88ff-f943dc337e8d" (UID: "1891d243-9edd-48aa-88ff-f943dc337e8d"). InnerVolumeSpecName "kube-api-access-bxv7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.911029 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1891d243-9edd-48aa-88ff-f943dc337e8d" (UID: "1891d243-9edd-48aa-88ff-f943dc337e8d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.922636 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-inventory" (OuterVolumeSpecName: "inventory") pod "1891d243-9edd-48aa-88ff-f943dc337e8d" (UID: "1891d243-9edd-48aa-88ff-f943dc337e8d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.980524 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxv7d\" (UniqueName: \"kubernetes.io/projected/1891d243-9edd-48aa-88ff-f943dc337e8d-kube-api-access-bxv7d\") on node \"crc\" DevicePath \"\"" Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.980560 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:32:40 crc kubenswrapper[4926]: I0312 18:32:40.980569 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1891d243-9edd-48aa-88ff-f943dc337e8d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.238672 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" event={"ID":"1891d243-9edd-48aa-88ff-f943dc337e8d","Type":"ContainerDied","Data":"3b48a2d4c059ac3596463fa9774e80dda754bd59c1f5280cd14ce2ca34cfd373"} Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.238713 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b48a2d4c059ac3596463fa9774e80dda754bd59c1f5280cd14ce2ca34cfd373" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.238782 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.334542 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679"] Mar 12 18:32:41 crc kubenswrapper[4926]: E0312 18:32:41.335263 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1891d243-9edd-48aa-88ff-f943dc337e8d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.335293 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1891d243-9edd-48aa-88ff-f943dc337e8d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.335548 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1891d243-9edd-48aa-88ff-f943dc337e8d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.336475 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.339078 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.339190 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.343015 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.343212 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.354860 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679"] Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.489492 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.489996 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ks6v\" (UniqueName: \"kubernetes.io/projected/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-kube-api-access-6ks6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.490093 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.591765 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.591872 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.591962 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ks6v\" (UniqueName: \"kubernetes.io/projected/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-kube-api-access-6ks6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.598327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.610578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.612234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ks6v\" (UniqueName: \"kubernetes.io/projected/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-kube-api-access-6ks6v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2d679\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:41 crc kubenswrapper[4926]: I0312 18:32:41.662660 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:32:42 crc kubenswrapper[4926]: I0312 18:32:42.230887 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679"] Mar 12 18:32:42 crc kubenswrapper[4926]: I0312 18:32:42.236215 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:32:42 crc kubenswrapper[4926]: I0312 18:32:42.257649 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" event={"ID":"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3","Type":"ContainerStarted","Data":"29b92aea004582f0f351798351f79cbd679e14449e74d5553f917bee426e56d9"} Mar 12 18:32:43 crc kubenswrapper[4926]: I0312 18:32:43.054848 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-4vtkq"] Mar 12 18:32:43 crc kubenswrapper[4926]: I0312 18:32:43.071870 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-4vtkq"] Mar 12 18:32:43 crc kubenswrapper[4926]: I0312 18:32:43.272654 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" event={"ID":"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3","Type":"ContainerStarted","Data":"3908783110cb82ca0bb6465283312cde6f24a0f3397f93570aca5ad317b5a1da"} Mar 12 18:32:43 crc kubenswrapper[4926]: I0312 18:32:43.300856 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" podStartSLOduration=1.8736497349999999 podStartE2EDuration="2.300819548s" podCreationTimestamp="2026-03-12 18:32:41 +0000 UTC" firstStartedPulling="2026-03-12 18:32:42.236005192 +0000 UTC m=+1802.604631515" lastFinishedPulling="2026-03-12 18:32:42.663175005 +0000 UTC m=+1803.031801328" observedRunningTime="2026-03-12 18:32:43.293477781 +0000 UTC m=+1803.662104134" watchObservedRunningTime="2026-03-12 18:32:43.300819548 +0000 UTC m=+1803.669445911" Mar 12 18:32:44 crc kubenswrapper[4926]: I0312 18:32:44.509222 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b6b9d3-0c3c-4dcc-b417-49d53269c39d" path="/var/lib/kubelet/pods/d7b6b9d3-0c3c-4dcc-b417-49d53269c39d/volumes" Mar 12 18:32:48 crc kubenswrapper[4926]: I0312 18:32:48.490693 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:32:48 crc kubenswrapper[4926]: E0312 18:32:48.491460 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:32:59 crc kubenswrapper[4926]: I0312 18:32:59.489835 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:32:59 crc kubenswrapper[4926]: E0312 18:32:59.490497 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:33:11 crc kubenswrapper[4926]: I0312 18:33:11.044364 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-vc5cr"] Mar 12 18:33:11 crc kubenswrapper[4926]: I0312 18:33:11.051425 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-vc5cr"] Mar 12 18:33:12 crc kubenswrapper[4926]: I0312 18:33:12.490780 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:33:12 crc kubenswrapper[4926]: E0312 18:33:12.491183 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:33:12 crc kubenswrapper[4926]: I0312 18:33:12.507389 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b" path="/var/lib/kubelet/pods/e1853a6c-a8bb-4d95-b02e-5b708c9d1b2b/volumes" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.162087 4926 scope.go:117] "RemoveContainer" containerID="d40939b357197123b7378525a59c3f91e7240cb30140873883a85a0672ad25de" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.222652 4926 scope.go:117] "RemoveContainer" containerID="f8a8e8e5fcda402042df2295b1c2cbd321aa245d9b7be946af5559093b56242a" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.273203 4926 scope.go:117] "RemoveContainer" containerID="7a2cc30c7cfe8ca6acf3b5029da578678f6497bf01186f5ff5a4caf50683498c" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.300485 4926 scope.go:117] "RemoveContainer" containerID="f7b35eaf50020882859b6b648c035a75d02c47868f155d967f93cf393c62f39e" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.355569 4926 scope.go:117] "RemoveContainer" containerID="1a95a2291b71efd71b4d69e1bff55bb10de0a0cbc11d599a6fbdc537cf2a1c21" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.405399 4926 scope.go:117] "RemoveContainer" containerID="7fd785662fcf1b6a3363e17fd1fb61ae49e121593b0c0273db1785eefe4c3db8" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.470732 4926 scope.go:117] "RemoveContainer" containerID="c3e155687eb67c8841a428e422ae982d9a3f9a577c50dd1606a11fd1ad93242c" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.493866 4926 scope.go:117] "RemoveContainer" containerID="ff84043d3f919db7adaecbb9d5f2f9199eabe47965d6033ec791be724e00a8f2" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.523521 4926 scope.go:117] "RemoveContainer" containerID="12643d715a7a25afe43a9b8c6f2f23a51e1343ba5e8cfc3caebf03243d5ac109" Mar 12 18:33:16 crc kubenswrapper[4926]: I0312 18:33:16.553729 4926 scope.go:117] "RemoveContainer" containerID="c4054fe5394aadede666252204c9b503dc03ddd88691e230f2deda1936fe7bc5" Mar 12 18:33:18 crc kubenswrapper[4926]: I0312 18:33:18.035332 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-txt96"] Mar 12 18:33:18 crc kubenswrapper[4926]: I0312 18:33:18.045292 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-txt96"] Mar 12 18:33:18 crc kubenswrapper[4926]: I0312 18:33:18.502925 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a0f2830-bf50-4195-9dad-d4d2c9529ee9" path="/var/lib/kubelet/pods/7a0f2830-bf50-4195-9dad-d4d2c9529ee9/volumes" Mar 12 18:33:24 crc kubenswrapper[4926]: I0312 18:33:24.490566 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:33:24 crc kubenswrapper[4926]: E0312 18:33:24.491300 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:33:25 crc kubenswrapper[4926]: I0312 18:33:25.035155 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-sgnbr"] Mar 12 18:33:25 crc kubenswrapper[4926]: I0312 18:33:25.044943 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-98lfj"] Mar 12 18:33:25 crc kubenswrapper[4926]: I0312 18:33:25.057020 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-sgnbr"] Mar 12 18:33:25 crc kubenswrapper[4926]: I0312 18:33:25.068062 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-98lfj"] Mar 12 18:33:26 crc kubenswrapper[4926]: I0312 18:33:26.501297 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af5704dd-cd13-4e5f-a77b-01266c63eeba" path="/var/lib/kubelet/pods/af5704dd-cd13-4e5f-a77b-01266c63eeba/volumes" Mar 12 18:33:26 crc kubenswrapper[4926]: I0312 18:33:26.502708 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e96fcb3d-2f9f-468d-bafa-060a9d1f1af6" path="/var/lib/kubelet/pods/e96fcb3d-2f9f-468d-bafa-060a9d1f1af6/volumes" Mar 12 18:33:30 crc kubenswrapper[4926]: I0312 18:33:30.774650 4926 generic.go:334] "Generic (PLEG): container finished" podID="2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3" containerID="3908783110cb82ca0bb6465283312cde6f24a0f3397f93570aca5ad317b5a1da" exitCode=0 Mar 12 18:33:30 crc kubenswrapper[4926]: I0312 18:33:30.774731 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" event={"ID":"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3","Type":"ContainerDied","Data":"3908783110cb82ca0bb6465283312cde6f24a0f3397f93570aca5ad317b5a1da"} Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.202473 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.303110 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-inventory\") pod \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.303240 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-ssh-key-openstack-edpm-ipam\") pod \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.303461 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ks6v\" (UniqueName: \"kubernetes.io/projected/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-kube-api-access-6ks6v\") pod \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\" (UID: \"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3\") " Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.322813 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-kube-api-access-6ks6v" (OuterVolumeSpecName: "kube-api-access-6ks6v") pod "2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3" (UID: "2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3"). InnerVolumeSpecName "kube-api-access-6ks6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.341128 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3" (UID: "2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.346657 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-inventory" (OuterVolumeSpecName: "inventory") pod "2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3" (UID: "2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.405332 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ks6v\" (UniqueName: \"kubernetes.io/projected/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-kube-api-access-6ks6v\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.405376 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.405389 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.808914 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" event={"ID":"2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3","Type":"ContainerDied","Data":"29b92aea004582f0f351798351f79cbd679e14449e74d5553f917bee426e56d9"} Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.809318 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29b92aea004582f0f351798351f79cbd679e14449e74d5553f917bee426e56d9" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.809401 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2d679" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.885510 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hckzq"] Mar 12 18:33:32 crc kubenswrapper[4926]: E0312 18:33:32.885885 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.885904 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.886083 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.886746 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.889043 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.889076 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.889683 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.890203 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:33:32 crc kubenswrapper[4926]: I0312 18:33:32.907905 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hckzq"] Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.018398 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.018473 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrgwg\" (UniqueName: \"kubernetes.io/projected/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-kube-api-access-jrgwg\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.018517 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.121523 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.121605 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrgwg\" (UniqueName: \"kubernetes.io/projected/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-kube-api-access-jrgwg\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.121658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.125076 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.127302 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.139897 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrgwg\" (UniqueName: \"kubernetes.io/projected/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-kube-api-access-jrgwg\") pod \"ssh-known-hosts-edpm-deployment-hckzq\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.202405 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.755783 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-hckzq"] Mar 12 18:33:33 crc kubenswrapper[4926]: I0312 18:33:33.818302 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" event={"ID":"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde","Type":"ContainerStarted","Data":"b85405818faa450f3f66fb64306238ecf32d8922ec0d96eecb12de32e6dcabc3"} Mar 12 18:33:34 crc kubenswrapper[4926]: I0312 18:33:34.828574 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" event={"ID":"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde","Type":"ContainerStarted","Data":"201651c1e9fe821527e37d1cb405f0d130718068782330f1ef3a8604311947b9"} Mar 12 18:33:34 crc kubenswrapper[4926]: I0312 18:33:34.856915 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" podStartSLOduration=2.408289462 podStartE2EDuration="2.85689196s" podCreationTimestamp="2026-03-12 18:33:32 +0000 UTC" firstStartedPulling="2026-03-12 18:33:33.763907509 +0000 UTC m=+1854.132533882" lastFinishedPulling="2026-03-12 18:33:34.212510037 +0000 UTC m=+1854.581136380" observedRunningTime="2026-03-12 18:33:34.846075714 +0000 UTC m=+1855.214702057" watchObservedRunningTime="2026-03-12 18:33:34.85689196 +0000 UTC m=+1855.225518303" Mar 12 18:33:35 crc kubenswrapper[4926]: I0312 18:33:35.496666 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:33:35 crc kubenswrapper[4926]: E0312 18:33:35.496983 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:33:38 crc kubenswrapper[4926]: I0312 18:33:38.044381 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-lgvzs"] Mar 12 18:33:38 crc kubenswrapper[4926]: I0312 18:33:38.057369 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-lgvzs"] Mar 12 18:33:38 crc kubenswrapper[4926]: I0312 18:33:38.501889 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dac4b5d6-fb31-4955-8679-db9d3ff63c10" path="/var/lib/kubelet/pods/dac4b5d6-fb31-4955-8679-db9d3ff63c10/volumes" Mar 12 18:33:41 crc kubenswrapper[4926]: I0312 18:33:41.907031 4926 generic.go:334] "Generic (PLEG): container finished" podID="d48eb883-81b1-4ad6-bbbe-f9c5a9779fde" containerID="201651c1e9fe821527e37d1cb405f0d130718068782330f1ef3a8604311947b9" exitCode=0 Mar 12 18:33:41 crc kubenswrapper[4926]: I0312 18:33:41.907084 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" event={"ID":"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde","Type":"ContainerDied","Data":"201651c1e9fe821527e37d1cb405f0d130718068782330f1ef3a8604311947b9"} Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.417649 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.540463 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-ssh-key-openstack-edpm-ipam\") pod \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.540605 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrgwg\" (UniqueName: \"kubernetes.io/projected/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-kube-api-access-jrgwg\") pod \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.541166 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-inventory-0\") pod \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\" (UID: \"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde\") " Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.551718 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-kube-api-access-jrgwg" (OuterVolumeSpecName: "kube-api-access-jrgwg") pod "d48eb883-81b1-4ad6-bbbe-f9c5a9779fde" (UID: "d48eb883-81b1-4ad6-bbbe-f9c5a9779fde"). InnerVolumeSpecName "kube-api-access-jrgwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.573718 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d48eb883-81b1-4ad6-bbbe-f9c5a9779fde" (UID: "d48eb883-81b1-4ad6-bbbe-f9c5a9779fde"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.591340 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d48eb883-81b1-4ad6-bbbe-f9c5a9779fde" (UID: "d48eb883-81b1-4ad6-bbbe-f9c5a9779fde"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.644884 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.644925 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrgwg\" (UniqueName: \"kubernetes.io/projected/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-kube-api-access-jrgwg\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.644939 4926 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d48eb883-81b1-4ad6-bbbe-f9c5a9779fde-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.926813 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" event={"ID":"d48eb883-81b1-4ad6-bbbe-f9c5a9779fde","Type":"ContainerDied","Data":"b85405818faa450f3f66fb64306238ecf32d8922ec0d96eecb12de32e6dcabc3"} Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.926865 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b85405818faa450f3f66fb64306238ecf32d8922ec0d96eecb12de32e6dcabc3" Mar 12 18:33:43 crc kubenswrapper[4926]: I0312 18:33:43.926929 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-hckzq" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.030953 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll"] Mar 12 18:33:44 crc kubenswrapper[4926]: E0312 18:33:44.031836 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d48eb883-81b1-4ad6-bbbe-f9c5a9779fde" containerName="ssh-known-hosts-edpm-deployment" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.031872 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d48eb883-81b1-4ad6-bbbe-f9c5a9779fde" containerName="ssh-known-hosts-edpm-deployment" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.032336 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d48eb883-81b1-4ad6-bbbe-f9c5a9779fde" containerName="ssh-known-hosts-edpm-deployment" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.033133 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.035133 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.035459 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.036141 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.049357 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll"] Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.052264 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.054327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.054684 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np6hd\" (UniqueName: \"kubernetes.io/projected/f12fab7b-7f88-441e-b230-551dc9ffa270-kube-api-access-np6hd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.054793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.155634 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.155798 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np6hd\" (UniqueName: \"kubernetes.io/projected/f12fab7b-7f88-441e-b230-551dc9ffa270-kube-api-access-np6hd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.155839 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.166655 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.166891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.175304 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np6hd\" (UniqueName: \"kubernetes.io/projected/f12fab7b-7f88-441e-b230-551dc9ffa270-kube-api-access-np6hd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-zxwll\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.357223 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.891154 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll"] Mar 12 18:33:44 crc kubenswrapper[4926]: I0312 18:33:44.936529 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" event={"ID":"f12fab7b-7f88-441e-b230-551dc9ffa270","Type":"ContainerStarted","Data":"2a4b919a20dc2bec583121219cdf56f1fee5a07381e675a44dfaa71d5cdaf38c"} Mar 12 18:33:45 crc kubenswrapper[4926]: I0312 18:33:45.961851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" event={"ID":"f12fab7b-7f88-441e-b230-551dc9ffa270","Type":"ContainerStarted","Data":"3eace6f424559836667819de65864f24f87afd1b5a74cbc8cf1b2289d0902d1f"} Mar 12 18:33:45 crc kubenswrapper[4926]: I0312 18:33:45.978667 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" podStartSLOduration=1.593961612 podStartE2EDuration="1.978645828s" podCreationTimestamp="2026-03-12 18:33:44 +0000 UTC" firstStartedPulling="2026-03-12 18:33:44.891205469 +0000 UTC m=+1865.259831802" lastFinishedPulling="2026-03-12 18:33:45.275889675 +0000 UTC m=+1865.644516018" observedRunningTime="2026-03-12 18:33:45.977358229 +0000 UTC m=+1866.345984602" watchObservedRunningTime="2026-03-12 18:33:45.978645828 +0000 UTC m=+1866.347272161" Mar 12 18:33:47 crc kubenswrapper[4926]: I0312 18:33:47.490322 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:33:47 crc kubenswrapper[4926]: E0312 18:33:47.490697 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:33:53 crc kubenswrapper[4926]: I0312 18:33:53.032276 4926 generic.go:334] "Generic (PLEG): container finished" podID="f12fab7b-7f88-441e-b230-551dc9ffa270" containerID="3eace6f424559836667819de65864f24f87afd1b5a74cbc8cf1b2289d0902d1f" exitCode=0 Mar 12 18:33:53 crc kubenswrapper[4926]: I0312 18:33:53.032383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" event={"ID":"f12fab7b-7f88-441e-b230-551dc9ffa270","Type":"ContainerDied","Data":"3eace6f424559836667819de65864f24f87afd1b5a74cbc8cf1b2289d0902d1f"} Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.519198 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.659632 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-inventory\") pod \"f12fab7b-7f88-441e-b230-551dc9ffa270\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.660126 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-ssh-key-openstack-edpm-ipam\") pod \"f12fab7b-7f88-441e-b230-551dc9ffa270\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.660258 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np6hd\" (UniqueName: \"kubernetes.io/projected/f12fab7b-7f88-441e-b230-551dc9ffa270-kube-api-access-np6hd\") pod \"f12fab7b-7f88-441e-b230-551dc9ffa270\" (UID: \"f12fab7b-7f88-441e-b230-551dc9ffa270\") " Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.667210 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12fab7b-7f88-441e-b230-551dc9ffa270-kube-api-access-np6hd" (OuterVolumeSpecName: "kube-api-access-np6hd") pod "f12fab7b-7f88-441e-b230-551dc9ffa270" (UID: "f12fab7b-7f88-441e-b230-551dc9ffa270"). InnerVolumeSpecName "kube-api-access-np6hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.691085 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-inventory" (OuterVolumeSpecName: "inventory") pod "f12fab7b-7f88-441e-b230-551dc9ffa270" (UID: "f12fab7b-7f88-441e-b230-551dc9ffa270"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.703577 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f12fab7b-7f88-441e-b230-551dc9ffa270" (UID: "f12fab7b-7f88-441e-b230-551dc9ffa270"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.764650 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.764797 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f12fab7b-7f88-441e-b230-551dc9ffa270-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:54 crc kubenswrapper[4926]: I0312 18:33:54.764899 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np6hd\" (UniqueName: \"kubernetes.io/projected/f12fab7b-7f88-441e-b230-551dc9ffa270-kube-api-access-np6hd\") on node \"crc\" DevicePath \"\"" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.056040 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" event={"ID":"f12fab7b-7f88-441e-b230-551dc9ffa270","Type":"ContainerDied","Data":"2a4b919a20dc2bec583121219cdf56f1fee5a07381e675a44dfaa71d5cdaf38c"} Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.056098 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a4b919a20dc2bec583121219cdf56f1fee5a07381e675a44dfaa71d5cdaf38c" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.056430 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-zxwll" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.149733 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc"] Mar 12 18:33:55 crc kubenswrapper[4926]: E0312 18:33:55.150523 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12fab7b-7f88-441e-b230-551dc9ffa270" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.150612 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12fab7b-7f88-441e-b230-551dc9ffa270" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.150893 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12fab7b-7f88-441e-b230-551dc9ffa270" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.151790 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.154921 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.157368 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.157638 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.157774 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-g65rs" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.168667 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc"] Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.274775 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8msml\" (UniqueName: \"kubernetes.io/projected/63c9d413-012b-47aa-a519-113685eb478c-kube-api-access-8msml\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.275359 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.275424 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.377630 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.377680 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.377766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8msml\" (UniqueName: \"kubernetes.io/projected/63c9d413-012b-47aa-a519-113685eb478c-kube-api-access-8msml\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.387212 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.387222 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.398327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8msml\" (UniqueName: \"kubernetes.io/projected/63c9d413-012b-47aa-a519-113685eb478c-kube-api-access-8msml\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:55 crc kubenswrapper[4926]: I0312 18:33:55.483988 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:33:56 crc kubenswrapper[4926]: W0312 18:33:56.026659 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63c9d413_012b_47aa_a519_113685eb478c.slice/crio-652a64dc7fb64ba5f266c774add42dc7aede4a98794574de6c44fc9a65426637 WatchSource:0}: Error finding container 652a64dc7fb64ba5f266c774add42dc7aede4a98794574de6c44fc9a65426637: Status 404 returned error can't find the container with id 652a64dc7fb64ba5f266c774add42dc7aede4a98794574de6c44fc9a65426637 Mar 12 18:33:56 crc kubenswrapper[4926]: I0312 18:33:56.030069 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc"] Mar 12 18:33:56 crc kubenswrapper[4926]: I0312 18:33:56.067365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" event={"ID":"63c9d413-012b-47aa-a519-113685eb478c","Type":"ContainerStarted","Data":"652a64dc7fb64ba5f266c774add42dc7aede4a98794574de6c44fc9a65426637"} Mar 12 18:33:57 crc kubenswrapper[4926]: I0312 18:33:57.079485 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" event={"ID":"63c9d413-012b-47aa-a519-113685eb478c","Type":"ContainerStarted","Data":"25268059ef3b3f8ab9572b4422f7b00bab2e017511cd7ff4bdb505fd3d2abecf"} Mar 12 18:33:57 crc kubenswrapper[4926]: I0312 18:33:57.100197 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" podStartSLOduration=1.559594038 podStartE2EDuration="2.10016892s" podCreationTimestamp="2026-03-12 18:33:55 +0000 UTC" firstStartedPulling="2026-03-12 18:33:56.028585173 +0000 UTC m=+1876.397211506" lastFinishedPulling="2026-03-12 18:33:56.569160055 +0000 UTC m=+1876.937786388" observedRunningTime="2026-03-12 18:33:57.099091657 +0000 UTC m=+1877.467717990" watchObservedRunningTime="2026-03-12 18:33:57.10016892 +0000 UTC m=+1877.468795273" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.132896 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555674-2b7bv"] Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.134743 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.137940 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.138497 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.138990 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.144090 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555674-2b7bv"] Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.283324 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xj22\" (UniqueName: \"kubernetes.io/projected/540a72de-64b2-41af-9caf-95895f11cb79-kube-api-access-7xj22\") pod \"auto-csr-approver-29555674-2b7bv\" (UID: \"540a72de-64b2-41af-9caf-95895f11cb79\") " pod="openshift-infra/auto-csr-approver-29555674-2b7bv" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.399487 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xj22\" (UniqueName: \"kubernetes.io/projected/540a72de-64b2-41af-9caf-95895f11cb79-kube-api-access-7xj22\") pod \"auto-csr-approver-29555674-2b7bv\" (UID: \"540a72de-64b2-41af-9caf-95895f11cb79\") " pod="openshift-infra/auto-csr-approver-29555674-2b7bv" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.427926 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xj22\" (UniqueName: \"kubernetes.io/projected/540a72de-64b2-41af-9caf-95895f11cb79-kube-api-access-7xj22\") pod \"auto-csr-approver-29555674-2b7bv\" (UID: \"540a72de-64b2-41af-9caf-95895f11cb79\") " pod="openshift-infra/auto-csr-approver-29555674-2b7bv" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.452906 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" Mar 12 18:34:00 crc kubenswrapper[4926]: I0312 18:34:00.887055 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555674-2b7bv"] Mar 12 18:34:00 crc kubenswrapper[4926]: W0312 18:34:00.897243 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod540a72de_64b2_41af_9caf_95895f11cb79.slice/crio-845fce3c4ea08b9cf4883148e9b4ea6682a491350bc2bfffb85900ad2ab21538 WatchSource:0}: Error finding container 845fce3c4ea08b9cf4883148e9b4ea6682a491350bc2bfffb85900ad2ab21538: Status 404 returned error can't find the container with id 845fce3c4ea08b9cf4883148e9b4ea6682a491350bc2bfffb85900ad2ab21538 Mar 12 18:34:01 crc kubenswrapper[4926]: I0312 18:34:01.121772 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" event={"ID":"540a72de-64b2-41af-9caf-95895f11cb79","Type":"ContainerStarted","Data":"845fce3c4ea08b9cf4883148e9b4ea6682a491350bc2bfffb85900ad2ab21538"} Mar 12 18:34:01 crc kubenswrapper[4926]: I0312 18:34:01.489572 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:34:01 crc kubenswrapper[4926]: E0312 18:34:01.489919 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:34:06 crc kubenswrapper[4926]: I0312 18:34:06.170812 4926 generic.go:334] "Generic (PLEG): container finished" podID="63c9d413-012b-47aa-a519-113685eb478c" containerID="25268059ef3b3f8ab9572b4422f7b00bab2e017511cd7ff4bdb505fd3d2abecf" exitCode=0 Mar 12 18:34:06 crc kubenswrapper[4926]: I0312 18:34:06.170969 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" event={"ID":"63c9d413-012b-47aa-a519-113685eb478c","Type":"ContainerDied","Data":"25268059ef3b3f8ab9572b4422f7b00bab2e017511cd7ff4bdb505fd3d2abecf"} Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.187759 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" event={"ID":"540a72de-64b2-41af-9caf-95895f11cb79","Type":"ContainerStarted","Data":"804c89c001a28f1eae59ba1c1cb327e0cc2c54749f5a07cd9cd07164cc1b86c4"} Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.216909 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" podStartSLOduration=1.356921442 podStartE2EDuration="7.21687683s" podCreationTimestamp="2026-03-12 18:34:00 +0000 UTC" firstStartedPulling="2026-03-12 18:34:00.899710811 +0000 UTC m=+1881.268337144" lastFinishedPulling="2026-03-12 18:34:06.759666159 +0000 UTC m=+1887.128292532" observedRunningTime="2026-03-12 18:34:07.205110269 +0000 UTC m=+1887.573736602" watchObservedRunningTime="2026-03-12 18:34:07.21687683 +0000 UTC m=+1887.585503203" Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.600684 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.743838 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-ssh-key-openstack-edpm-ipam\") pod \"63c9d413-012b-47aa-a519-113685eb478c\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.744019 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-inventory\") pod \"63c9d413-012b-47aa-a519-113685eb478c\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.744086 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8msml\" (UniqueName: \"kubernetes.io/projected/63c9d413-012b-47aa-a519-113685eb478c-kube-api-access-8msml\") pod \"63c9d413-012b-47aa-a519-113685eb478c\" (UID: \"63c9d413-012b-47aa-a519-113685eb478c\") " Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.751539 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c9d413-012b-47aa-a519-113685eb478c-kube-api-access-8msml" (OuterVolumeSpecName: "kube-api-access-8msml") pod "63c9d413-012b-47aa-a519-113685eb478c" (UID: "63c9d413-012b-47aa-a519-113685eb478c"). InnerVolumeSpecName "kube-api-access-8msml". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.775952 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-inventory" (OuterVolumeSpecName: "inventory") pod "63c9d413-012b-47aa-a519-113685eb478c" (UID: "63c9d413-012b-47aa-a519-113685eb478c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.779774 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "63c9d413-012b-47aa-a519-113685eb478c" (UID: "63c9d413-012b-47aa-a519-113685eb478c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.847466 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.847521 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63c9d413-012b-47aa-a519-113685eb478c-inventory\") on node \"crc\" DevicePath \"\"" Mar 12 18:34:07 crc kubenswrapper[4926]: I0312 18:34:07.847540 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8msml\" (UniqueName: \"kubernetes.io/projected/63c9d413-012b-47aa-a519-113685eb478c-kube-api-access-8msml\") on node \"crc\" DevicePath \"\"" Mar 12 18:34:08 crc kubenswrapper[4926]: I0312 18:34:08.205163 4926 generic.go:334] "Generic (PLEG): container finished" podID="540a72de-64b2-41af-9caf-95895f11cb79" containerID="804c89c001a28f1eae59ba1c1cb327e0cc2c54749f5a07cd9cd07164cc1b86c4" exitCode=0 Mar 12 18:34:08 crc kubenswrapper[4926]: I0312 18:34:08.205236 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" event={"ID":"540a72de-64b2-41af-9caf-95895f11cb79","Type":"ContainerDied","Data":"804c89c001a28f1eae59ba1c1cb327e0cc2c54749f5a07cd9cd07164cc1b86c4"} Mar 12 18:34:08 crc kubenswrapper[4926]: I0312 18:34:08.213053 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" event={"ID":"63c9d413-012b-47aa-a519-113685eb478c","Type":"ContainerDied","Data":"652a64dc7fb64ba5f266c774add42dc7aede4a98794574de6c44fc9a65426637"} Mar 12 18:34:08 crc kubenswrapper[4926]: I0312 18:34:08.213124 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="652a64dc7fb64ba5f266c774add42dc7aede4a98794574de6c44fc9a65426637" Mar 12 18:34:08 crc kubenswrapper[4926]: I0312 18:34:08.213188 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc" Mar 12 18:34:09 crc kubenswrapper[4926]: I0312 18:34:09.575990 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" Mar 12 18:34:09 crc kubenswrapper[4926]: I0312 18:34:09.683595 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xj22\" (UniqueName: \"kubernetes.io/projected/540a72de-64b2-41af-9caf-95895f11cb79-kube-api-access-7xj22\") pod \"540a72de-64b2-41af-9caf-95895f11cb79\" (UID: \"540a72de-64b2-41af-9caf-95895f11cb79\") " Mar 12 18:34:09 crc kubenswrapper[4926]: I0312 18:34:09.691003 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540a72de-64b2-41af-9caf-95895f11cb79-kube-api-access-7xj22" (OuterVolumeSpecName: "kube-api-access-7xj22") pod "540a72de-64b2-41af-9caf-95895f11cb79" (UID: "540a72de-64b2-41af-9caf-95895f11cb79"). InnerVolumeSpecName "kube-api-access-7xj22". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:34:09 crc kubenswrapper[4926]: I0312 18:34:09.786260 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xj22\" (UniqueName: \"kubernetes.io/projected/540a72de-64b2-41af-9caf-95895f11cb79-kube-api-access-7xj22\") on node \"crc\" DevicePath \"\"" Mar 12 18:34:10 crc kubenswrapper[4926]: I0312 18:34:10.242259 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" event={"ID":"540a72de-64b2-41af-9caf-95895f11cb79","Type":"ContainerDied","Data":"845fce3c4ea08b9cf4883148e9b4ea6682a491350bc2bfffb85900ad2ab21538"} Mar 12 18:34:10 crc kubenswrapper[4926]: I0312 18:34:10.242648 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="845fce3c4ea08b9cf4883148e9b4ea6682a491350bc2bfffb85900ad2ab21538" Mar 12 18:34:10 crc kubenswrapper[4926]: I0312 18:34:10.242335 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555674-2b7bv" Mar 12 18:34:10 crc kubenswrapper[4926]: I0312 18:34:10.306614 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555668-xnfj4"] Mar 12 18:34:10 crc kubenswrapper[4926]: I0312 18:34:10.317165 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555668-xnfj4"] Mar 12 18:34:10 crc kubenswrapper[4926]: I0312 18:34:10.504221 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddbd98cc-c135-485c-9b18-c31f19321a38" path="/var/lib/kubelet/pods/ddbd98cc-c135-485c-9b18-c31f19321a38/volumes" Mar 12 18:34:12 crc kubenswrapper[4926]: I0312 18:34:12.490638 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:34:12 crc kubenswrapper[4926]: E0312 18:34:12.491329 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:34:16 crc kubenswrapper[4926]: I0312 18:34:16.774466 4926 scope.go:117] "RemoveContainer" containerID="d333e02f3e64bf137ca7da16020fb743166cb48ce41f070ab0d454452957be39" Mar 12 18:34:16 crc kubenswrapper[4926]: I0312 18:34:16.815110 4926 scope.go:117] "RemoveContainer" containerID="1b2fbc6d93b78d1252c9b5b62fe658eafa93cdb86d8acefaa37b868dbfa19698" Mar 12 18:34:16 crc kubenswrapper[4926]: I0312 18:34:16.867311 4926 scope.go:117] "RemoveContainer" containerID="96adbf85f267b1793c500e8c152eb5c14d98f7080edf6f6f116872d3c431cb0a" Mar 12 18:34:16 crc kubenswrapper[4926]: I0312 18:34:16.905138 4926 scope.go:117] "RemoveContainer" containerID="74c22411866d45d292c4082e0d4be48c52ad4683049047a45e2de32d9da3478f" Mar 12 18:34:16 crc kubenswrapper[4926]: I0312 18:34:16.950516 4926 scope.go:117] "RemoveContainer" containerID="5b07bfa58e0c68e62027f67884fd04fdf4f157c10b82889fbed4fe1cd06870d8" Mar 12 18:34:22 crc kubenswrapper[4926]: I0312 18:34:22.049393 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-382f-account-create-update-7sn9z"] Mar 12 18:34:22 crc kubenswrapper[4926]: I0312 18:34:22.062519 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-55g54"] Mar 12 18:34:22 crc kubenswrapper[4926]: I0312 18:34:22.070001 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-382f-account-create-update-7sn9z"] Mar 12 18:34:22 crc kubenswrapper[4926]: I0312 18:34:22.077410 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-55g54"] Mar 12 18:34:22 crc kubenswrapper[4926]: I0312 18:34:22.509722 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cb8f27f-336b-4acc-8165-0c53c1643084" path="/var/lib/kubelet/pods/4cb8f27f-336b-4acc-8165-0c53c1643084/volumes" Mar 12 18:34:22 crc kubenswrapper[4926]: I0312 18:34:22.510998 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe0c0d35-e463-4b5c-a15f-cdfc808c498c" path="/var/lib/kubelet/pods/fe0c0d35-e463-4b5c-a15f-cdfc808c498c/volumes" Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.038036 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-0630-account-create-update-knx9n"] Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.046494 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f45b-account-create-update-rt9j9"] Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.057626 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-t4krn"] Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.067681 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-cctxg"] Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.075197 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-0630-account-create-update-knx9n"] Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.082491 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f45b-account-create-update-rt9j9"] Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.090327 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-cctxg"] Mar 12 18:34:23 crc kubenswrapper[4926]: I0312 18:34:23.097818 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-t4krn"] Mar 12 18:34:24 crc kubenswrapper[4926]: I0312 18:34:24.511387 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15db9185-0441-431d-98db-5701f9b244be" path="/var/lib/kubelet/pods/15db9185-0441-431d-98db-5701f9b244be/volumes" Mar 12 18:34:24 crc kubenswrapper[4926]: I0312 18:34:24.513625 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a2618b-1e0a-4246-b63c-582d2fe2f847" path="/var/lib/kubelet/pods/29a2618b-1e0a-4246-b63c-582d2fe2f847/volumes" Mar 12 18:34:24 crc kubenswrapper[4926]: I0312 18:34:24.515223 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b33ff126-1b0c-47f0-a4c1-18e1297fa81d" path="/var/lib/kubelet/pods/b33ff126-1b0c-47f0-a4c1-18e1297fa81d/volumes" Mar 12 18:34:24 crc kubenswrapper[4926]: I0312 18:34:24.516590 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc4b3e47-ab9e-46c0-be37-9d2e94721be6" path="/var/lib/kubelet/pods/dc4b3e47-ab9e-46c0-be37-9d2e94721be6/volumes" Mar 12 18:34:25 crc kubenswrapper[4926]: I0312 18:34:25.490154 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:34:25 crc kubenswrapper[4926]: E0312 18:34:25.490755 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:34:39 crc kubenswrapper[4926]: I0312 18:34:39.490795 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:34:39 crc kubenswrapper[4926]: E0312 18:34:39.492100 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:34:48 crc kubenswrapper[4926]: I0312 18:34:48.067638 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cbqv8"] Mar 12 18:34:48 crc kubenswrapper[4926]: I0312 18:34:48.081171 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cbqv8"] Mar 12 18:34:48 crc kubenswrapper[4926]: I0312 18:34:48.503957 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57" path="/var/lib/kubelet/pods/436f0bee-a5ab-4d4f-bbaa-3c4c568f2a57/volumes" Mar 12 18:34:52 crc kubenswrapper[4926]: I0312 18:34:52.491022 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:34:52 crc kubenswrapper[4926]: E0312 18:34:52.492166 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.900822 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nz8z6/must-gather-x4gfw"] Mar 12 18:34:54 crc kubenswrapper[4926]: E0312 18:34:54.902507 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="540a72de-64b2-41af-9caf-95895f11cb79" containerName="oc" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.902613 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="540a72de-64b2-41af-9caf-95895f11cb79" containerName="oc" Mar 12 18:34:54 crc kubenswrapper[4926]: E0312 18:34:54.902732 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c9d413-012b-47aa-a519-113685eb478c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.902819 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c9d413-012b-47aa-a519-113685eb478c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.903128 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="540a72de-64b2-41af-9caf-95895f11cb79" containerName="oc" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.903227 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c9d413-012b-47aa-a519-113685eb478c" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.904509 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.906562 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nz8z6"/"openshift-service-ca.crt" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.906886 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nz8z6"/"kube-root-ca.crt" Mar 12 18:34:54 crc kubenswrapper[4926]: I0312 18:34:54.932965 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nz8z6/must-gather-x4gfw"] Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.008086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2562de93-d8c3-4055-8ab2-3c55b4f3c830-must-gather-output\") pod \"must-gather-x4gfw\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.008165 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w44tw\" (UniqueName: \"kubernetes.io/projected/2562de93-d8c3-4055-8ab2-3c55b4f3c830-kube-api-access-w44tw\") pod \"must-gather-x4gfw\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.110861 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2562de93-d8c3-4055-8ab2-3c55b4f3c830-must-gather-output\") pod \"must-gather-x4gfw\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.111229 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2562de93-d8c3-4055-8ab2-3c55b4f3c830-must-gather-output\") pod \"must-gather-x4gfw\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.111425 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w44tw\" (UniqueName: \"kubernetes.io/projected/2562de93-d8c3-4055-8ab2-3c55b4f3c830-kube-api-access-w44tw\") pod \"must-gather-x4gfw\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.131084 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w44tw\" (UniqueName: \"kubernetes.io/projected/2562de93-d8c3-4055-8ab2-3c55b4f3c830-kube-api-access-w44tw\") pod \"must-gather-x4gfw\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.224904 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.700547 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nz8z6/must-gather-x4gfw"] Mar 12 18:34:55 crc kubenswrapper[4926]: I0312 18:34:55.755092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" event={"ID":"2562de93-d8c3-4055-8ab2-3c55b4f3c830","Type":"ContainerStarted","Data":"93a53026295ef16f089cdb9826fdb5dbdb0b51d38d9f2aedd4f23b4040bfc0b8"} Mar 12 18:35:02 crc kubenswrapper[4926]: I0312 18:35:02.847730 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" event={"ID":"2562de93-d8c3-4055-8ab2-3c55b4f3c830","Type":"ContainerStarted","Data":"9983eed33b5073c82867ffe643c25901fb02264d34aa9779308efc3e4497884a"} Mar 12 18:35:02 crc kubenswrapper[4926]: I0312 18:35:02.848384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" event={"ID":"2562de93-d8c3-4055-8ab2-3c55b4f3c830","Type":"ContainerStarted","Data":"4dbea05e4f0fb083c128a04cb1592adbe2b7259461b2a690428955b1ed545c5d"} Mar 12 18:35:02 crc kubenswrapper[4926]: I0312 18:35:02.891207 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" podStartSLOduration=2.944702807 podStartE2EDuration="8.891185854s" podCreationTimestamp="2026-03-12 18:34:54 +0000 UTC" firstStartedPulling="2026-03-12 18:34:55.70678955 +0000 UTC m=+1936.075415913" lastFinishedPulling="2026-03-12 18:35:01.653272627 +0000 UTC m=+1942.021898960" observedRunningTime="2026-03-12 18:35:02.881807678 +0000 UTC m=+1943.250434041" watchObservedRunningTime="2026-03-12 18:35:02.891185854 +0000 UTC m=+1943.259812187" Mar 12 18:35:04 crc kubenswrapper[4926]: I0312 18:35:04.490303 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:35:04 crc kubenswrapper[4926]: E0312 18:35:04.490887 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.304269 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nz8z6/crc-debug-ss92r"] Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.306197 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.309509 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-nz8z6"/"default-dockercfg-9j8mc" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.448095 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h879n\" (UniqueName: \"kubernetes.io/projected/5381bdf5-d86c-4da5-ac97-b5a3a1997005-kube-api-access-h879n\") pod \"crc-debug-ss92r\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.448582 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5381bdf5-d86c-4da5-ac97-b5a3a1997005-host\") pod \"crc-debug-ss92r\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.551064 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5381bdf5-d86c-4da5-ac97-b5a3a1997005-host\") pod \"crc-debug-ss92r\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.551132 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h879n\" (UniqueName: \"kubernetes.io/projected/5381bdf5-d86c-4da5-ac97-b5a3a1997005-kube-api-access-h879n\") pod \"crc-debug-ss92r\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.551268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5381bdf5-d86c-4da5-ac97-b5a3a1997005-host\") pod \"crc-debug-ss92r\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.580069 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h879n\" (UniqueName: \"kubernetes.io/projected/5381bdf5-d86c-4da5-ac97-b5a3a1997005-kube-api-access-h879n\") pod \"crc-debug-ss92r\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.626112 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:06 crc kubenswrapper[4926]: W0312 18:35:06.671557 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5381bdf5_d86c_4da5_ac97_b5a3a1997005.slice/crio-8650f457341f9acefdfbaa4223783f6fe0533785330f983f51126ef0fa194e76 WatchSource:0}: Error finding container 8650f457341f9acefdfbaa4223783f6fe0533785330f983f51126ef0fa194e76: Status 404 returned error can't find the container with id 8650f457341f9acefdfbaa4223783f6fe0533785330f983f51126ef0fa194e76 Mar 12 18:35:06 crc kubenswrapper[4926]: I0312 18:35:06.890196 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" event={"ID":"5381bdf5-d86c-4da5-ac97-b5a3a1997005","Type":"ContainerStarted","Data":"8650f457341f9acefdfbaa4223783f6fe0533785330f983f51126ef0fa194e76"} Mar 12 18:35:10 crc kubenswrapper[4926]: I0312 18:35:10.032958 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-88dpb"] Mar 12 18:35:10 crc kubenswrapper[4926]: I0312 18:35:10.050022 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-88dpb"] Mar 12 18:35:10 crc kubenswrapper[4926]: I0312 18:35:10.500792 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d30718d9-b986-4490-9e43-56eaa459aeb5" path="/var/lib/kubelet/pods/d30718d9-b986-4490-9e43-56eaa459aeb5/volumes" Mar 12 18:35:11 crc kubenswrapper[4926]: I0312 18:35:11.041413 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2wcs"] Mar 12 18:35:11 crc kubenswrapper[4926]: I0312 18:35:11.048049 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-x2wcs"] Mar 12 18:35:12 crc kubenswrapper[4926]: I0312 18:35:12.503279 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b478ba05-2155-4b13-a58a-002be25403d0" path="/var/lib/kubelet/pods/b478ba05-2155-4b13-a58a-002be25403d0/volumes" Mar 12 18:35:17 crc kubenswrapper[4926]: I0312 18:35:17.071467 4926 scope.go:117] "RemoveContainer" containerID="a9bf2ae52889d181443948b724481f92343c2befd9093bc8cc132900dface592" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.586013 4926 scope.go:117] "RemoveContainer" containerID="6afc876661934d69ed6bb90885f853f8b5249b4339a13cab5b59071dc578e401" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.628825 4926 scope.go:117] "RemoveContainer" containerID="976b5bff6bf1beb72cdab8caedbb157ee653f67334ae07735bc5919349b4ae10" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.678712 4926 scope.go:117] "RemoveContainer" containerID="683758cbf77fef46bca458acf7d282b0867f1aa1a065fbc9925548f2b1c73316" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.740477 4926 scope.go:117] "RemoveContainer" containerID="2f4d726cd118a123d7b32175fc4e520ff68fb0c877630e02cc5267fde94f47a0" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.836986 4926 scope.go:117] "RemoveContainer" containerID="14cbad167526774b9aa42516a514a2cffabd952a439286d5c8bfe698bbd81e8b" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.888310 4926 scope.go:117] "RemoveContainer" containerID="8476342456babac729c7086b287eb9b60e559eb560bedd1d78d42784e15f0ff2" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.906693 4926 scope.go:117] "RemoveContainer" containerID="f80b93ed7cbf010c51c1bbf4a01b83fdc0dc54adf44c99408ef3f7476eea541b" Mar 12 18:35:18 crc kubenswrapper[4926]: I0312 18:35:18.930729 4926 scope.go:117] "RemoveContainer" containerID="b0665b4cdb9e47cd07d74cad2d92a7eeb8aadf990846b2f50cc7f0a3a493d8dd" Mar 12 18:35:19 crc kubenswrapper[4926]: I0312 18:35:19.046966 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" event={"ID":"5381bdf5-d86c-4da5-ac97-b5a3a1997005","Type":"ContainerStarted","Data":"3b7cae5f88beda72e15102b000ae8d919d1219da1f42e5b5be4f546f7fa97f6c"} Mar 12 18:35:19 crc kubenswrapper[4926]: I0312 18:35:19.063952 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" podStartSLOduration=1.05588493 podStartE2EDuration="13.063930328s" podCreationTimestamp="2026-03-12 18:35:06 +0000 UTC" firstStartedPulling="2026-03-12 18:35:06.67388103 +0000 UTC m=+1947.042507363" lastFinishedPulling="2026-03-12 18:35:18.681926428 +0000 UTC m=+1959.050552761" observedRunningTime="2026-03-12 18:35:19.061777631 +0000 UTC m=+1959.430403974" watchObservedRunningTime="2026-03-12 18:35:19.063930328 +0000 UTC m=+1959.432556661" Mar 12 18:35:19 crc kubenswrapper[4926]: I0312 18:35:19.489987 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:35:19 crc kubenswrapper[4926]: E0312 18:35:19.490207 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:35:25 crc kubenswrapper[4926]: I0312 18:35:25.033818 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-vwzff"] Mar 12 18:35:25 crc kubenswrapper[4926]: I0312 18:35:25.045977 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-vwzff"] Mar 12 18:35:26 crc kubenswrapper[4926]: I0312 18:35:26.502827 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a36d0b-fd08-4c93-846c-688c71055113" path="/var/lib/kubelet/pods/76a36d0b-fd08-4c93-846c-688c71055113/volumes" Mar 12 18:35:33 crc kubenswrapper[4926]: I0312 18:35:33.202452 4926 generic.go:334] "Generic (PLEG): container finished" podID="5381bdf5-d86c-4da5-ac97-b5a3a1997005" containerID="3b7cae5f88beda72e15102b000ae8d919d1219da1f42e5b5be4f546f7fa97f6c" exitCode=0 Mar 12 18:35:33 crc kubenswrapper[4926]: I0312 18:35:33.202506 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" event={"ID":"5381bdf5-d86c-4da5-ac97-b5a3a1997005","Type":"ContainerDied","Data":"3b7cae5f88beda72e15102b000ae8d919d1219da1f42e5b5be4f546f7fa97f6c"} Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.335788 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.368501 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nz8z6/crc-debug-ss92r"] Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.373740 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nz8z6/crc-debug-ss92r"] Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.384197 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h879n\" (UniqueName: \"kubernetes.io/projected/5381bdf5-d86c-4da5-ac97-b5a3a1997005-kube-api-access-h879n\") pod \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.384478 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5381bdf5-d86c-4da5-ac97-b5a3a1997005-host\") pod \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\" (UID: \"5381bdf5-d86c-4da5-ac97-b5a3a1997005\") " Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.384658 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5381bdf5-d86c-4da5-ac97-b5a3a1997005-host" (OuterVolumeSpecName: "host") pod "5381bdf5-d86c-4da5-ac97-b5a3a1997005" (UID: "5381bdf5-d86c-4da5-ac97-b5a3a1997005"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.384936 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5381bdf5-d86c-4da5-ac97-b5a3a1997005-host\") on node \"crc\" DevicePath \"\"" Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.388699 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5381bdf5-d86c-4da5-ac97-b5a3a1997005-kube-api-access-h879n" (OuterVolumeSpecName: "kube-api-access-h879n") pod "5381bdf5-d86c-4da5-ac97-b5a3a1997005" (UID: "5381bdf5-d86c-4da5-ac97-b5a3a1997005"). InnerVolumeSpecName "kube-api-access-h879n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.486233 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h879n\" (UniqueName: \"kubernetes.io/projected/5381bdf5-d86c-4da5-ac97-b5a3a1997005-kube-api-access-h879n\") on node \"crc\" DevicePath \"\"" Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.489605 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:35:34 crc kubenswrapper[4926]: E0312 18:35:34.489902 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:35:34 crc kubenswrapper[4926]: I0312 18:35:34.499779 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5381bdf5-d86c-4da5-ac97-b5a3a1997005" path="/var/lib/kubelet/pods/5381bdf5-d86c-4da5-ac97-b5a3a1997005/volumes" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.220294 4926 scope.go:117] "RemoveContainer" containerID="3b7cae5f88beda72e15102b000ae8d919d1219da1f42e5b5be4f546f7fa97f6c" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.220395 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-ss92r" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.574155 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nz8z6/crc-debug-6jp9x"] Mar 12 18:35:35 crc kubenswrapper[4926]: E0312 18:35:35.574513 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5381bdf5-d86c-4da5-ac97-b5a3a1997005" containerName="container-00" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.574525 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="5381bdf5-d86c-4da5-ac97-b5a3a1997005" containerName="container-00" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.574690 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="5381bdf5-d86c-4da5-ac97-b5a3a1997005" containerName="container-00" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.575240 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.578286 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-nz8z6"/"default-dockercfg-9j8mc" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.606358 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/23ffd21c-99e8-47f0-945a-70fb2da71885-host\") pod \"crc-debug-6jp9x\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.606506 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhlff\" (UniqueName: \"kubernetes.io/projected/23ffd21c-99e8-47f0-945a-70fb2da71885-kube-api-access-lhlff\") pod \"crc-debug-6jp9x\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.715889 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/23ffd21c-99e8-47f0-945a-70fb2da71885-host\") pod \"crc-debug-6jp9x\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.716007 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhlff\" (UniqueName: \"kubernetes.io/projected/23ffd21c-99e8-47f0-945a-70fb2da71885-kube-api-access-lhlff\") pod \"crc-debug-6jp9x\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.716050 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/23ffd21c-99e8-47f0-945a-70fb2da71885-host\") pod \"crc-debug-6jp9x\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.734194 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhlff\" (UniqueName: \"kubernetes.io/projected/23ffd21c-99e8-47f0-945a-70fb2da71885-kube-api-access-lhlff\") pod \"crc-debug-6jp9x\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:35 crc kubenswrapper[4926]: I0312 18:35:35.891709 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:36 crc kubenswrapper[4926]: I0312 18:35:36.235707 4926 generic.go:334] "Generic (PLEG): container finished" podID="23ffd21c-99e8-47f0-945a-70fb2da71885" containerID="c1196ac145f33bde3cf7ef54fd8462b66ce61b83f03e51afe39dafbe2872f728" exitCode=1 Mar 12 18:35:36 crc kubenswrapper[4926]: I0312 18:35:36.235835 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" event={"ID":"23ffd21c-99e8-47f0-945a-70fb2da71885","Type":"ContainerDied","Data":"c1196ac145f33bde3cf7ef54fd8462b66ce61b83f03e51afe39dafbe2872f728"} Mar 12 18:35:36 crc kubenswrapper[4926]: I0312 18:35:36.236028 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" event={"ID":"23ffd21c-99e8-47f0-945a-70fb2da71885","Type":"ContainerStarted","Data":"fd73f0549d58a377ffca80b1bd0fbac707795e9afe2863c99dc4b3bf9df893a3"} Mar 12 18:35:36 crc kubenswrapper[4926]: I0312 18:35:36.279646 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nz8z6/crc-debug-6jp9x"] Mar 12 18:35:36 crc kubenswrapper[4926]: I0312 18:35:36.288422 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nz8z6/crc-debug-6jp9x"] Mar 12 18:35:36 crc kubenswrapper[4926]: E0312 18:35:36.342697 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23ffd21c_99e8_47f0_945a_70fb2da71885.slice/crio-conmon-c1196ac145f33bde3cf7ef54fd8462b66ce61b83f03e51afe39dafbe2872f728.scope\": RecentStats: unable to find data in memory cache]" Mar 12 18:35:37 crc kubenswrapper[4926]: I0312 18:35:37.347713 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:37 crc kubenswrapper[4926]: I0312 18:35:37.449031 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/23ffd21c-99e8-47f0-945a-70fb2da71885-host\") pod \"23ffd21c-99e8-47f0-945a-70fb2da71885\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " Mar 12 18:35:37 crc kubenswrapper[4926]: I0312 18:35:37.449126 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23ffd21c-99e8-47f0-945a-70fb2da71885-host" (OuterVolumeSpecName: "host") pod "23ffd21c-99e8-47f0-945a-70fb2da71885" (UID: "23ffd21c-99e8-47f0-945a-70fb2da71885"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 18:35:37 crc kubenswrapper[4926]: I0312 18:35:37.449652 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhlff\" (UniqueName: \"kubernetes.io/projected/23ffd21c-99e8-47f0-945a-70fb2da71885-kube-api-access-lhlff\") pod \"23ffd21c-99e8-47f0-945a-70fb2da71885\" (UID: \"23ffd21c-99e8-47f0-945a-70fb2da71885\") " Mar 12 18:35:37 crc kubenswrapper[4926]: I0312 18:35:37.450121 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/23ffd21c-99e8-47f0-945a-70fb2da71885-host\") on node \"crc\" DevicePath \"\"" Mar 12 18:35:37 crc kubenswrapper[4926]: I0312 18:35:37.455263 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23ffd21c-99e8-47f0-945a-70fb2da71885-kube-api-access-lhlff" (OuterVolumeSpecName: "kube-api-access-lhlff") pod "23ffd21c-99e8-47f0-945a-70fb2da71885" (UID: "23ffd21c-99e8-47f0-945a-70fb2da71885"). InnerVolumeSpecName "kube-api-access-lhlff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:35:37 crc kubenswrapper[4926]: I0312 18:35:37.551578 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhlff\" (UniqueName: \"kubernetes.io/projected/23ffd21c-99e8-47f0-945a-70fb2da71885-kube-api-access-lhlff\") on node \"crc\" DevicePath \"\"" Mar 12 18:35:38 crc kubenswrapper[4926]: I0312 18:35:38.257583 4926 scope.go:117] "RemoveContainer" containerID="c1196ac145f33bde3cf7ef54fd8462b66ce61b83f03e51afe39dafbe2872f728" Mar 12 18:35:38 crc kubenswrapper[4926]: I0312 18:35:38.257704 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/crc-debug-6jp9x" Mar 12 18:35:38 crc kubenswrapper[4926]: I0312 18:35:38.501790 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23ffd21c-99e8-47f0-945a-70fb2da71885" path="/var/lib/kubelet/pods/23ffd21c-99e8-47f0-945a-70fb2da71885/volumes" Mar 12 18:35:46 crc kubenswrapper[4926]: I0312 18:35:46.490644 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:35:46 crc kubenswrapper[4926]: E0312 18:35:46.491512 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:35:59 crc kubenswrapper[4926]: I0312 18:35:59.490481 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:35:59 crc kubenswrapper[4926]: E0312 18:35:59.491204 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.149017 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555676-5d5p5"] Mar 12 18:36:00 crc kubenswrapper[4926]: E0312 18:36:00.149490 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23ffd21c-99e8-47f0-945a-70fb2da71885" containerName="container-00" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.149519 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="23ffd21c-99e8-47f0-945a-70fb2da71885" containerName="container-00" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.149732 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="23ffd21c-99e8-47f0-945a-70fb2da71885" containerName="container-00" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.151530 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.154512 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.155083 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.156753 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.187597 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555676-5d5p5"] Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.277975 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6tqr\" (UniqueName: \"kubernetes.io/projected/f1b045ea-1fff-43b3-9a66-48dd361e9f33-kube-api-access-v6tqr\") pod \"auto-csr-approver-29555676-5d5p5\" (UID: \"f1b045ea-1fff-43b3-9a66-48dd361e9f33\") " pod="openshift-infra/auto-csr-approver-29555676-5d5p5" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.380825 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6tqr\" (UniqueName: \"kubernetes.io/projected/f1b045ea-1fff-43b3-9a66-48dd361e9f33-kube-api-access-v6tqr\") pod \"auto-csr-approver-29555676-5d5p5\" (UID: \"f1b045ea-1fff-43b3-9a66-48dd361e9f33\") " pod="openshift-infra/auto-csr-approver-29555676-5d5p5" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.404043 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6tqr\" (UniqueName: \"kubernetes.io/projected/f1b045ea-1fff-43b3-9a66-48dd361e9f33-kube-api-access-v6tqr\") pod \"auto-csr-approver-29555676-5d5p5\" (UID: \"f1b045ea-1fff-43b3-9a66-48dd361e9f33\") " pod="openshift-infra/auto-csr-approver-29555676-5d5p5" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.473709 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" Mar 12 18:36:00 crc kubenswrapper[4926]: I0312 18:36:00.931014 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555676-5d5p5"] Mar 12 18:36:01 crc kubenswrapper[4926]: I0312 18:36:01.469480 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" event={"ID":"f1b045ea-1fff-43b3-9a66-48dd361e9f33","Type":"ContainerStarted","Data":"89b811e4e20a47fa80d513af23713e88bc0b9d51b9526b50771abdfdfd97db80"} Mar 12 18:36:02 crc kubenswrapper[4926]: I0312 18:36:02.480781 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" event={"ID":"f1b045ea-1fff-43b3-9a66-48dd361e9f33","Type":"ContainerStarted","Data":"cfe5fa51265434fa1414bed949d3e915d15d11d193d0cf363993fcfab40b91d4"} Mar 12 18:36:02 crc kubenswrapper[4926]: I0312 18:36:02.497510 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" podStartSLOduration=1.337521948 podStartE2EDuration="2.497494259s" podCreationTimestamp="2026-03-12 18:36:00 +0000 UTC" firstStartedPulling="2026-03-12 18:36:00.936511179 +0000 UTC m=+2001.305137532" lastFinishedPulling="2026-03-12 18:36:02.09648351 +0000 UTC m=+2002.465109843" observedRunningTime="2026-03-12 18:36:02.496195318 +0000 UTC m=+2002.864821651" watchObservedRunningTime="2026-03-12 18:36:02.497494259 +0000 UTC m=+2002.866120592" Mar 12 18:36:03 crc kubenswrapper[4926]: I0312 18:36:03.489786 4926 generic.go:334] "Generic (PLEG): container finished" podID="f1b045ea-1fff-43b3-9a66-48dd361e9f33" containerID="cfe5fa51265434fa1414bed949d3e915d15d11d193d0cf363993fcfab40b91d4" exitCode=0 Mar 12 18:36:03 crc kubenswrapper[4926]: I0312 18:36:03.489909 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" event={"ID":"f1b045ea-1fff-43b3-9a66-48dd361e9f33","Type":"ContainerDied","Data":"cfe5fa51265434fa1414bed949d3e915d15d11d193d0cf363993fcfab40b91d4"} Mar 12 18:36:04 crc kubenswrapper[4926]: I0312 18:36:04.904655 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" Mar 12 18:36:04 crc kubenswrapper[4926]: I0312 18:36:04.995987 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6tqr\" (UniqueName: \"kubernetes.io/projected/f1b045ea-1fff-43b3-9a66-48dd361e9f33-kube-api-access-v6tqr\") pod \"f1b045ea-1fff-43b3-9a66-48dd361e9f33\" (UID: \"f1b045ea-1fff-43b3-9a66-48dd361e9f33\") " Mar 12 18:36:05 crc kubenswrapper[4926]: I0312 18:36:05.015686 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1b045ea-1fff-43b3-9a66-48dd361e9f33-kube-api-access-v6tqr" (OuterVolumeSpecName: "kube-api-access-v6tqr") pod "f1b045ea-1fff-43b3-9a66-48dd361e9f33" (UID: "f1b045ea-1fff-43b3-9a66-48dd361e9f33"). InnerVolumeSpecName "kube-api-access-v6tqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:36:05 crc kubenswrapper[4926]: I0312 18:36:05.098561 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6tqr\" (UniqueName: \"kubernetes.io/projected/f1b045ea-1fff-43b3-9a66-48dd361e9f33-kube-api-access-v6tqr\") on node \"crc\" DevicePath \"\"" Mar 12 18:36:05 crc kubenswrapper[4926]: I0312 18:36:05.508107 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" event={"ID":"f1b045ea-1fff-43b3-9a66-48dd361e9f33","Type":"ContainerDied","Data":"89b811e4e20a47fa80d513af23713e88bc0b9d51b9526b50771abdfdfd97db80"} Mar 12 18:36:05 crc kubenswrapper[4926]: I0312 18:36:05.508163 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89b811e4e20a47fa80d513af23713e88bc0b9d51b9526b50771abdfdfd97db80" Mar 12 18:36:05 crc kubenswrapper[4926]: I0312 18:36:05.508250 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555676-5d5p5" Mar 12 18:36:05 crc kubenswrapper[4926]: I0312 18:36:05.575988 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555670-v8lqv"] Mar 12 18:36:05 crc kubenswrapper[4926]: I0312 18:36:05.587781 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555670-v8lqv"] Mar 12 18:36:06 crc kubenswrapper[4926]: I0312 18:36:06.504943 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daa768ee-1644-40fa-8f52-790b28d9bb74" path="/var/lib/kubelet/pods/daa768ee-1644-40fa-8f52-790b28d9bb74/volumes" Mar 12 18:36:11 crc kubenswrapper[4926]: I0312 18:36:11.490048 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:36:11 crc kubenswrapper[4926]: E0312 18:36:11.490899 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:36:15 crc kubenswrapper[4926]: I0312 18:36:15.786452 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6fd6447dc6-7dg85_32458d94-2727-4718-a842-f20e68b6a0dd/barbican-api/0.log" Mar 12 18:36:15 crc kubenswrapper[4926]: I0312 18:36:15.955992 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6fd6447dc6-7dg85_32458d94-2727-4718-a842-f20e68b6a0dd/barbican-api-log/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.000377 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74668b896d-mctls_4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92/barbican-keystone-listener/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.024517 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74668b896d-mctls_4cce8ba3-9eb9-4ed4-a3bf-a6787fdaae92/barbican-keystone-listener-log/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.189904 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fddddf9f9-kbftb_08cef0ec-16bc-4b64-95f6-e0d8f22fa00e/barbican-worker/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.205296 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-fddddf9f9-kbftb_08cef0ec-16bc-4b64-95f6-e0d8f22fa00e/barbican-worker-log/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.383996 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f983b88e-aba3-4d49-bbd4-4db5eef5266c/ceilometer-central-agent/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.446922 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-88k6z_75a3208b-42f5-412e-a503-ac328f7d9967/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.474103 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f983b88e-aba3-4d49-bbd4-4db5eef5266c/ceilometer-notification-agent/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.562788 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f983b88e-aba3-4d49-bbd4-4db5eef5266c/proxy-httpd/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.626336 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f983b88e-aba3-4d49-bbd4-4db5eef5266c/sg-core/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.656562 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-fm77w_1891d243-9edd-48aa-88ff-f943dc337e8d/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.837083 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_edd14509-9e82-40a2-aea4-c6ad4250be05/cinder-api/0.log" Mar 12 18:36:16 crc kubenswrapper[4926]: I0312 18:36:16.846394 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_edd14509-9e82-40a2-aea4-c6ad4250be05/cinder-api-log/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.021706 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7b586d98-12e7-4814-82c6-6724e1b35a77/cinder-scheduler/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.053693 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_7b586d98-12e7-4814-82c6-6724e1b35a77/probe/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.177900 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-h2dvq_eab3d0c2-5edc-4657-928f-52a87de2293a/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.240338 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2d679_2fd86e0a-1dc5-4f07-96fb-e9d15b32cca3/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.434531 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5fc8b56cf-6cbwx_6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69/init/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.598537 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5fc8b56cf-6cbwx_6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69/dnsmasq-dns/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.624531 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5fc8b56cf-6cbwx_6f42a0ec-9b43-4a24-b5b2-89cd0f3abe69/init/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.643193 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0749ce08-7fa5-48fe-9248-ac3a6699ef57/glance-httpd/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.822836 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_1e75a0d1-4272-43f3-b1b8-3cfe57e0141d/glance-httpd/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.836599 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0749ce08-7fa5-48fe-9248-ac3a6699ef57/glance-log/0.log" Mar 12 18:36:17 crc kubenswrapper[4926]: I0312 18:36:17.868040 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_1e75a0d1-4272-43f3-b1b8-3cfe57e0141d/glance-log/0.log" Mar 12 18:36:18 crc kubenswrapper[4926]: I0312 18:36:18.126540 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c6848d8cd-cq57n_a1ae8f23-3518-430a-bbcf-e7be0cb8282e/horizon/0.log" Mar 12 18:36:18 crc kubenswrapper[4926]: I0312 18:36:18.146115 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5c6848d8cd-cq57n_a1ae8f23-3518-430a-bbcf-e7be0cb8282e/horizon-log/0.log" Mar 12 18:36:18 crc kubenswrapper[4926]: I0312 18:36:18.289739 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-njwrr_7f4f99b4-e1a5-4b29-b5e4-c4fcd240ca46/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:18 crc kubenswrapper[4926]: I0312 18:36:18.426238 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5b667c464b-fk8sc_054d27f0-5c9b-4e59-98b3-e05609c3b257/keystone-api/0.log" Mar 12 18:36:18 crc kubenswrapper[4926]: I0312 18:36:18.496329 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_30b7fb7c-fcab-4551-8284-c0dab53beb21/kube-state-metrics/0.log" Mar 12 18:36:18 crc kubenswrapper[4926]: I0312 18:36:18.739379 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7fc6496dbc-7qwrw_bf1502ef-50a4-45e0-b193-a6e25abccb32/neutron-api/0.log" Mar 12 18:36:18 crc kubenswrapper[4926]: I0312 18:36:18.852128 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7fc6496dbc-7qwrw_bf1502ef-50a4-45e0-b193-a6e25abccb32/neutron-httpd/0.log" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.109915 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_65b09111-c033-45e3-97d3-cd755e1a79ab/nova-api-api/0.log" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.145641 4926 scope.go:117] "RemoveContainer" containerID="ddcfcec936ed5e6887603a245e014b4f9336991c54bce882c0b7657f1d5a318b" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.184364 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_65b09111-c033-45e3-97d3-cd755e1a79ab/nova-api-log/0.log" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.199128 4926 scope.go:117] "RemoveContainer" containerID="799ca781836933eed5f381ea7d6b69d2905b6867e2a978eb9f38bd2b549ed6cb" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.450592 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_397b3426-1e6c-468a-8d6d-562e70944d9d/nova-cell0-conductor-conductor/0.log" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.504397 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_9696ed5e-263b-457b-a771-9f7a7e27dbed/nova-cell1-conductor-conductor/0.log" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.724166 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5990afe5-179c-401d-99eb-58b27e2bfc9e/nova-cell1-novncproxy-novncproxy/0.log" Mar 12 18:36:19 crc kubenswrapper[4926]: I0312 18:36:19.788553 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8065d406-f127-4c57-b603-e7e6afeb3731/nova-metadata-log/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.045133 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_6d9c2983-b118-4677-ba18-20531d4223ad/nova-scheduler-scheduler/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.070149 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8065d406-f127-4c57-b603-e7e6afeb3731/nova-metadata-metadata/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.187512 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_94d38c16-a6c9-44ed-a49e-398dc34b92ce/mysql-bootstrap/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.334094 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_94d38c16-a6c9-44ed-a49e-398dc34b92ce/mysql-bootstrap/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.416898 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1ee00086-3c8a-4f3a-a5d5-9590715a8b95/mysql-bootstrap/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.427172 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_94d38c16-a6c9-44ed-a49e-398dc34b92ce/galera/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.562161 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1ee00086-3c8a-4f3a-a5d5-9590715a8b95/mysql-bootstrap/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.644280 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_1ee00086-3c8a-4f3a-a5d5-9590715a8b95/galera/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.658263 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_8ac368f3-42fb-4f4a-ba68-1686386b017e/openstackclient/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.838533 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mszqk_b4a45633-8ac7-497f-a8d7-2b7a3dad35bc/openstack-network-exporter/0.log" Mar 12 18:36:20 crc kubenswrapper[4926]: I0312 18:36:20.942600 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6znbv_30dfa384-92a5-49cf-9793-60478855264f/ovsdb-server-init/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.046120 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6znbv_30dfa384-92a5-49cf-9793-60478855264f/ovs-vswitchd/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.094391 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6znbv_30dfa384-92a5-49cf-9793-60478855264f/ovsdb-server/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.116048 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6znbv_30dfa384-92a5-49cf-9793-60478855264f/ovsdb-server-init/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.288372 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-sfwpr_8f19fbb7-ea3b-437a-a634-498e6a593ef6/ovn-controller/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.360284 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ffbc2f0-cbec-43d4-9907-a95af037ae1b/openstack-network-exporter/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.382462 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3ffbc2f0-cbec-43d4-9907-a95af037ae1b/ovn-northd/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.530683 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c3288572-a9dc-4f96-8535-b05f6f22855b/ovsdbserver-nb/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.531628 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c3288572-a9dc-4f96-8535-b05f6f22855b/openstack-network-exporter/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.719993 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8926ccac-b553-4e37-bbcb-96e3b00c1cab/ovsdbserver-sb/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.806151 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_8926ccac-b553-4e37-bbcb-96e3b00c1cab/openstack-network-exporter/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.861390 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6fdd856968-nmxxn_0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81/placement-api/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.941193 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6fdd856968-nmxxn_0c666ce1-e9f9-480f-a2eb-91c2cb9e7d81/placement-log/0.log" Mar 12 18:36:21 crc kubenswrapper[4926]: I0312 18:36:21.984949 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b9b82f03-7ac1-4805-858b-708760b4e476/setup-container/0.log" Mar 12 18:36:22 crc kubenswrapper[4926]: I0312 18:36:22.300900 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b9b82f03-7ac1-4805-858b-708760b4e476/rabbitmq/0.log" Mar 12 18:36:22 crc kubenswrapper[4926]: I0312 18:36:22.326684 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b9b82f03-7ac1-4805-858b-708760b4e476/setup-container/0.log" Mar 12 18:36:22 crc kubenswrapper[4926]: I0312 18:36:22.370284 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9aba0434-585e-4355-8019-1612400b2350/setup-container/0.log" Mar 12 18:36:22 crc kubenswrapper[4926]: I0312 18:36:22.530165 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9aba0434-585e-4355-8019-1612400b2350/rabbitmq/0.log" Mar 12 18:36:22 crc kubenswrapper[4926]: I0312 18:36:22.578239 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_9aba0434-585e-4355-8019-1612400b2350/setup-container/0.log" Mar 12 18:36:22 crc kubenswrapper[4926]: I0312 18:36:22.595826 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-5c2tc_63c9d413-012b-47aa-a519-113685eb478c/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:22 crc kubenswrapper[4926]: I0312 18:36:22.988964 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g45rs_58e2a40e-84ba-43fe-8087-ec2e917a2b34/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.035273 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-zxwll_f12fab7b-7f88-441e-b230-551dc9ffa270/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.214389 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-hckzq_d48eb883-81b1-4ad6-bbbe-f9c5a9779fde/ssh-known-hosts-edpm-deployment/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.298949 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-bbdc94bc7-pmtmm_09eb7b3b-c5af-4625-8f1a-83766550711c/proxy-httpd/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.474632 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-bbdc94bc7-pmtmm_09eb7b3b-c5af-4625-8f1a-83766550711c/proxy-server/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.489615 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:36:23 crc kubenswrapper[4926]: E0312 18:36:23.489884 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.538235 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bn6gf_9e8d5f15-0f8f-4e98-a9fc-cf0925c0990f/swift-ring-rebalance/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.664400 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/account-auditor/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.702831 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/account-reaper/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.752167 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/account-server/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.806450 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/account-replicator/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.868650 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/container-auditor/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.933337 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/container-replicator/0.log" Mar 12 18:36:23 crc kubenswrapper[4926]: I0312 18:36:23.970942 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/container-updater/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.006807 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/container-server/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.084341 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/object-auditor/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.142879 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/object-expirer/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.181750 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/object-replicator/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.260307 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/object-server/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.331046 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/object-updater/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.366882 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/rsync/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.419393 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_57853681-32de-4475-9c7d-3f9708fe7d91/swift-recon-cron/0.log" Mar 12 18:36:24 crc kubenswrapper[4926]: I0312 18:36:24.570899 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-55zdj_96a114de-73e3-4088-902e-c3d1fcaaa3ad/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 12 18:36:28 crc kubenswrapper[4926]: I0312 18:36:28.525408 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_69083379-a7d7-4876-9955-497420eab579/memcached/0.log" Mar 12 18:36:35 crc kubenswrapper[4926]: I0312 18:36:35.489680 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:36:35 crc kubenswrapper[4926]: E0312 18:36:35.490446 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:36:47 crc kubenswrapper[4926]: I0312 18:36:47.246379 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt_1a47463c-c539-4a5c-a3fd-c09feef2de67/util/0.log" Mar 12 18:36:47 crc kubenswrapper[4926]: I0312 18:36:47.343898 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt_1a47463c-c539-4a5c-a3fd-c09feef2de67/util/0.log" Mar 12 18:36:47 crc kubenswrapper[4926]: I0312 18:36:47.425361 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt_1a47463c-c539-4a5c-a3fd-c09feef2de67/pull/0.log" Mar 12 18:36:47 crc kubenswrapper[4926]: I0312 18:36:47.448111 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt_1a47463c-c539-4a5c-a3fd-c09feef2de67/pull/0.log" Mar 12 18:36:47 crc kubenswrapper[4926]: I0312 18:36:47.600562 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt_1a47463c-c539-4a5c-a3fd-c09feef2de67/util/0.log" Mar 12 18:36:47 crc kubenswrapper[4926]: I0312 18:36:47.627583 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt_1a47463c-c539-4a5c-a3fd-c09feef2de67/pull/0.log" Mar 12 18:36:47 crc kubenswrapper[4926]: I0312 18:36:47.647333 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0a9cf9d9b218433ba8d40c603f7c72428a887310de9e6eb6a71674eecedmpzt_1a47463c-c539-4a5c-a3fd-c09feef2de67/extract/0.log" Mar 12 18:36:48 crc kubenswrapper[4926]: I0312 18:36:48.038501 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-clngf_4800992f-cfad-4a1c-94e5-79427f88c002/manager/0.log" Mar 12 18:36:48 crc kubenswrapper[4926]: I0312 18:36:48.372542 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-m6sxh_b776be98-1352-43c6-8ee8-e31076b7d12b/manager/0.log" Mar 12 18:36:48 crc kubenswrapper[4926]: I0312 18:36:48.633975 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-525z5_fea71415-42ac-4e77-ba9c-25170ccece27/manager/0.log" Mar 12 18:36:48 crc kubenswrapper[4926]: I0312 18:36:48.830924 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-p76mx_065fe73a-651c-4cd3-b8d7-135617c51bbd/manager/0.log" Mar 12 18:36:49 crc kubenswrapper[4926]: I0312 18:36:49.176886 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-zlsrd_f112cb87-7454-41fa-a1e1-381d79f86247/manager/0.log" Mar 12 18:36:49 crc kubenswrapper[4926]: I0312 18:36:49.356030 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-7q8xr_a41da562-a119-4785-95d0-eaf0970a99f4/manager/0.log" Mar 12 18:36:49 crc kubenswrapper[4926]: I0312 18:36:49.508534 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-kzl7p_fd35525c-7b73-49d1-a36c-c49d3bf933eb/manager/0.log" Mar 12 18:36:49 crc kubenswrapper[4926]: I0312 18:36:49.700245 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-5f6c4_c62baaa0-d72b-4240-9dba-858bdf61d1b3/manager/0.log" Mar 12 18:36:49 crc kubenswrapper[4926]: I0312 18:36:49.770879 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-f96w7_e38cf931-bbd6-4b47-bdf6-8a514d17d3d7/manager/0.log" Mar 12 18:36:50 crc kubenswrapper[4926]: I0312 18:36:50.010518 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-c58lh_ccd465e8-6811-4865-9602-3dea8144cc01/manager/0.log" Mar 12 18:36:50 crc kubenswrapper[4926]: I0312 18:36:50.331419 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-njrjb_6374b396-00ef-4aca-ac07-fd46982f23f1/manager/0.log" Mar 12 18:36:50 crc kubenswrapper[4926]: I0312 18:36:50.466140 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-569cc54c5-dzj5w_2e44c177-b87d-4ff6-80ca-672477fe9e94/manager/0.log" Mar 12 18:36:50 crc kubenswrapper[4926]: I0312 18:36:50.497307 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:36:50 crc kubenswrapper[4926]: E0312 18:36:50.497614 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:36:50 crc kubenswrapper[4926]: I0312 18:36:50.596607 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-cv97b_464725b8-2734-43a4-a232-5db9bafed311/manager/0.log" Mar 12 18:36:50 crc kubenswrapper[4926]: I0312 18:36:50.798519 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7cxtpk_5d4dea90-1696-4195-a0a0-71a3c9f3e328/manager/0.log" Mar 12 18:36:51 crc kubenswrapper[4926]: I0312 18:36:51.498068 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-66f4595798-c276m_fe4ee666-a3c1-44c5-a07b-5ca8438e0482/operator/0.log" Mar 12 18:36:51 crc kubenswrapper[4926]: I0312 18:36:51.674361 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-hwc2c_66ae05ba-2fc2-4915-9899-083a49295427/registry-server/0.log" Mar 12 18:36:51 crc kubenswrapper[4926]: I0312 18:36:51.853836 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-xr747_14346445-95be-488f-858c-44bf5b45c656/manager/0.log" Mar 12 18:36:51 crc kubenswrapper[4926]: I0312 18:36:51.981187 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-tpnlt_e3baa344-9dd8-48ed-8b6a-60ff9fbc181a/manager/0.log" Mar 12 18:36:52 crc kubenswrapper[4926]: I0312 18:36:52.147397 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-wqzk9_fc6165e6-3d8f-4ddd-b6ae-a1307f2c6b3b/operator/0.log" Mar 12 18:36:52 crc kubenswrapper[4926]: I0312 18:36:52.363125 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-r2cxk_309d4a2a-f738-4d2d-a28e-f361f762f997/manager/0.log" Mar 12 18:36:52 crc kubenswrapper[4926]: I0312 18:36:52.579752 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-z87mn_45ddf7b5-1b73-473c-9da8-c35d9a4e0ddd/manager/0.log" Mar 12 18:36:52 crc kubenswrapper[4926]: I0312 18:36:52.626316 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-25jrt_682f3a0f-7437-455c-99e1-8b7cdb03328a/manager/0.log" Mar 12 18:36:52 crc kubenswrapper[4926]: I0312 18:36:52.839737 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-cn7fq_80f26a3f-ab1e-49b1-8843-6674d948a5cd/manager/0.log" Mar 12 18:36:52 crc kubenswrapper[4926]: I0312 18:36:52.908650 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55976db558-2kgv6_495cec8d-a262-4d55-9ee5-6eebb10b6765/manager/0.log" Mar 12 18:36:54 crc kubenswrapper[4926]: I0312 18:36:54.284789 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-ssmx7_e7b3fba0-ddaa-4cef-9df6-0683a92475cf/manager/0.log" Mar 12 18:37:01 crc kubenswrapper[4926]: I0312 18:37:01.258600 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-bbdc94bc7-pmtmm" podUID="09eb7b3b-c5af-4625-8f1a-83766550711c" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 12 18:37:02 crc kubenswrapper[4926]: I0312 18:37:02.490423 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:37:03 crc kubenswrapper[4926]: I0312 18:37:03.029422 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"2201cfc89c392b2d9df343a46350dae7f7620675e753c895f182c00c5eb9467c"} Mar 12 18:37:14 crc kubenswrapper[4926]: I0312 18:37:14.068688 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-sgl9z_5d073c88-4608-4594-9feb-f1093455368d/control-plane-machine-set-operator/0.log" Mar 12 18:37:14 crc kubenswrapper[4926]: I0312 18:37:14.276473 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bfn44_49e5e304-df7c-434b-8b17-f520e9bb7d52/kube-rbac-proxy/0.log" Mar 12 18:37:14 crc kubenswrapper[4926]: I0312 18:37:14.302575 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-bfn44_49e5e304-df7c-434b-8b17-f520e9bb7d52/machine-api-operator/0.log" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.516739 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ngnmz"] Mar 12 18:37:24 crc kubenswrapper[4926]: E0312 18:37:24.521040 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b045ea-1fff-43b3-9a66-48dd361e9f33" containerName="oc" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.521205 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b045ea-1fff-43b3-9a66-48dd361e9f33" containerName="oc" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.521525 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1b045ea-1fff-43b3-9a66-48dd361e9f33" containerName="oc" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.523246 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.535974 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ngnmz"] Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.670244 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcbv8\" (UniqueName: \"kubernetes.io/projected/305b4cae-bc27-4c5d-8142-042446a753af-kube-api-access-mcbv8\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.670295 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-utilities\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.670399 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-catalog-content\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.772073 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-catalog-content\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.772201 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcbv8\" (UniqueName: \"kubernetes.io/projected/305b4cae-bc27-4c5d-8142-042446a753af-kube-api-access-mcbv8\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.772232 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-utilities\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.772634 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-catalog-content\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.772793 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-utilities\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.810140 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcbv8\" (UniqueName: \"kubernetes.io/projected/305b4cae-bc27-4c5d-8142-042446a753af-kube-api-access-mcbv8\") pod \"redhat-operators-ngnmz\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:24 crc kubenswrapper[4926]: I0312 18:37:24.845873 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:25 crc kubenswrapper[4926]: I0312 18:37:25.316955 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ngnmz"] Mar 12 18:37:25 crc kubenswrapper[4926]: W0312 18:37:25.323377 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod305b4cae_bc27_4c5d_8142_042446a753af.slice/crio-27c46b747fc3b972108b73d6610b46b16d2ab44f98be0213b6dafd0b1655a885 WatchSource:0}: Error finding container 27c46b747fc3b972108b73d6610b46b16d2ab44f98be0213b6dafd0b1655a885: Status 404 returned error can't find the container with id 27c46b747fc3b972108b73d6610b46b16d2ab44f98be0213b6dafd0b1655a885 Mar 12 18:37:26 crc kubenswrapper[4926]: I0312 18:37:26.281900 4926 generic.go:334] "Generic (PLEG): container finished" podID="305b4cae-bc27-4c5d-8142-042446a753af" containerID="3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91" exitCode=0 Mar 12 18:37:26 crc kubenswrapper[4926]: I0312 18:37:26.282008 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ngnmz" event={"ID":"305b4cae-bc27-4c5d-8142-042446a753af","Type":"ContainerDied","Data":"3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91"} Mar 12 18:37:26 crc kubenswrapper[4926]: I0312 18:37:26.282316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ngnmz" event={"ID":"305b4cae-bc27-4c5d-8142-042446a753af","Type":"ContainerStarted","Data":"27c46b747fc3b972108b73d6610b46b16d2ab44f98be0213b6dafd0b1655a885"} Mar 12 18:37:27 crc kubenswrapper[4926]: I0312 18:37:27.291477 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ngnmz" event={"ID":"305b4cae-bc27-4c5d-8142-042446a753af","Type":"ContainerStarted","Data":"45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3"} Mar 12 18:37:27 crc kubenswrapper[4926]: I0312 18:37:27.693693 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-7pf7h_04933adf-efe6-4d54-8575-cc5c4069ea9a/cert-manager-controller/0.log" Mar 12 18:37:27 crc kubenswrapper[4926]: I0312 18:37:27.837972 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-mbhfc_eaa86db1-fe85-4b00-b8e0-c61cb013f52d/cert-manager-cainjector/0.log" Mar 12 18:37:27 crc kubenswrapper[4926]: I0312 18:37:27.888796 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-p8p7c_ef14eb59-d30a-437c-80d1-70513a544b2d/cert-manager-webhook/0.log" Mar 12 18:37:29 crc kubenswrapper[4926]: I0312 18:37:29.313123 4926 generic.go:334] "Generic (PLEG): container finished" podID="305b4cae-bc27-4c5d-8142-042446a753af" containerID="45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3" exitCode=0 Mar 12 18:37:29 crc kubenswrapper[4926]: I0312 18:37:29.313186 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ngnmz" event={"ID":"305b4cae-bc27-4c5d-8142-042446a753af","Type":"ContainerDied","Data":"45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3"} Mar 12 18:37:31 crc kubenswrapper[4926]: I0312 18:37:31.349316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ngnmz" event={"ID":"305b4cae-bc27-4c5d-8142-042446a753af","Type":"ContainerStarted","Data":"74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f"} Mar 12 18:37:31 crc kubenswrapper[4926]: I0312 18:37:31.368742 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ngnmz" podStartSLOduration=3.175810541 podStartE2EDuration="7.368727456s" podCreationTimestamp="2026-03-12 18:37:24 +0000 UTC" firstStartedPulling="2026-03-12 18:37:26.283793936 +0000 UTC m=+2086.652420269" lastFinishedPulling="2026-03-12 18:37:30.476710851 +0000 UTC m=+2090.845337184" observedRunningTime="2026-03-12 18:37:31.368277451 +0000 UTC m=+2091.736903784" watchObservedRunningTime="2026-03-12 18:37:31.368727456 +0000 UTC m=+2091.737353789" Mar 12 18:37:34 crc kubenswrapper[4926]: I0312 18:37:34.846488 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:34 crc kubenswrapper[4926]: I0312 18:37:34.848133 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:37:35 crc kubenswrapper[4926]: I0312 18:37:35.892223 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ngnmz" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="registry-server" probeResult="failure" output=< Mar 12 18:37:35 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:37:35 crc kubenswrapper[4926]: > Mar 12 18:37:41 crc kubenswrapper[4926]: I0312 18:37:41.235951 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-67hvb_9e731d72-a0a8-46a4-af0b-5dce65f29dd1/nmstate-console-plugin/0.log" Mar 12 18:37:41 crc kubenswrapper[4926]: I0312 18:37:41.389465 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-z52rq_9db224ca-f640-4756-9c80-afe7ff63dcbe/nmstate-handler/0.log" Mar 12 18:37:41 crc kubenswrapper[4926]: I0312 18:37:41.411272 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-rn7dj_cf97acff-2650-4e84-ab30-d10f9bd70ef4/kube-rbac-proxy/0.log" Mar 12 18:37:41 crc kubenswrapper[4926]: I0312 18:37:41.582177 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-rn7dj_cf97acff-2650-4e84-ab30-d10f9bd70ef4/nmstate-metrics/0.log" Mar 12 18:37:41 crc kubenswrapper[4926]: I0312 18:37:41.597584 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-wnkdv_7247f6a4-e62d-44e3-b91c-1117fca5c960/nmstate-operator/0.log" Mar 12 18:37:41 crc kubenswrapper[4926]: I0312 18:37:41.765538 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-h5ll8_bdc3504f-bf1b-4b71-aaf2-45e24e41a84e/nmstate-webhook/0.log" Mar 12 18:37:45 crc kubenswrapper[4926]: I0312 18:37:45.892489 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ngnmz" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="registry-server" probeResult="failure" output=< Mar 12 18:37:45 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:37:45 crc kubenswrapper[4926]: > Mar 12 18:37:55 crc kubenswrapper[4926]: I0312 18:37:55.909664 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ngnmz" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="registry-server" probeResult="failure" output=< Mar 12 18:37:55 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Mar 12 18:37:55 crc kubenswrapper[4926]: > Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.140567 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555678-f2ldc"] Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.142018 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555678-f2ldc" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.143704 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.144075 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.144738 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.151961 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555678-f2ldc"] Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.175268 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgcwh\" (UniqueName: \"kubernetes.io/projected/6e31677c-4037-4cbc-9d34-5487431793e9-kube-api-access-rgcwh\") pod \"auto-csr-approver-29555678-f2ldc\" (UID: \"6e31677c-4037-4cbc-9d34-5487431793e9\") " pod="openshift-infra/auto-csr-approver-29555678-f2ldc" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.277625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgcwh\" (UniqueName: \"kubernetes.io/projected/6e31677c-4037-4cbc-9d34-5487431793e9-kube-api-access-rgcwh\") pod \"auto-csr-approver-29555678-f2ldc\" (UID: \"6e31677c-4037-4cbc-9d34-5487431793e9\") " pod="openshift-infra/auto-csr-approver-29555678-f2ldc" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.302378 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgcwh\" (UniqueName: \"kubernetes.io/projected/6e31677c-4037-4cbc-9d34-5487431793e9-kube-api-access-rgcwh\") pod \"auto-csr-approver-29555678-f2ldc\" (UID: \"6e31677c-4037-4cbc-9d34-5487431793e9\") " pod="openshift-infra/auto-csr-approver-29555678-f2ldc" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.462450 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555678-f2ldc" Mar 12 18:38:00 crc kubenswrapper[4926]: I0312 18:38:00.999274 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555678-f2ldc"] Mar 12 18:38:01 crc kubenswrapper[4926]: I0312 18:38:01.007479 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:38:01 crc kubenswrapper[4926]: I0312 18:38:01.633293 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555678-f2ldc" event={"ID":"6e31677c-4037-4cbc-9d34-5487431793e9","Type":"ContainerStarted","Data":"45c44d1933bdfb1f99d54cecba14b4c68a52a137566e9c1b97d4dacb2c549a18"} Mar 12 18:38:02 crc kubenswrapper[4926]: I0312 18:38:02.648989 4926 generic.go:334] "Generic (PLEG): container finished" podID="6e31677c-4037-4cbc-9d34-5487431793e9" containerID="d822df10a89aa630f25e85c9dfeda5e7eb5b0c4e3ca5846540130f6ca7031958" exitCode=0 Mar 12 18:38:02 crc kubenswrapper[4926]: I0312 18:38:02.649197 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555678-f2ldc" event={"ID":"6e31677c-4037-4cbc-9d34-5487431793e9","Type":"ContainerDied","Data":"d822df10a89aa630f25e85c9dfeda5e7eb5b0c4e3ca5846540130f6ca7031958"} Mar 12 18:38:03 crc kubenswrapper[4926]: I0312 18:38:03.988464 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555678-f2ldc" Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.047392 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgcwh\" (UniqueName: \"kubernetes.io/projected/6e31677c-4037-4cbc-9d34-5487431793e9-kube-api-access-rgcwh\") pod \"6e31677c-4037-4cbc-9d34-5487431793e9\" (UID: \"6e31677c-4037-4cbc-9d34-5487431793e9\") " Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.053055 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e31677c-4037-4cbc-9d34-5487431793e9-kube-api-access-rgcwh" (OuterVolumeSpecName: "kube-api-access-rgcwh") pod "6e31677c-4037-4cbc-9d34-5487431793e9" (UID: "6e31677c-4037-4cbc-9d34-5487431793e9"). InnerVolumeSpecName "kube-api-access-rgcwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.149250 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgcwh\" (UniqueName: \"kubernetes.io/projected/6e31677c-4037-4cbc-9d34-5487431793e9-kube-api-access-rgcwh\") on node \"crc\" DevicePath \"\"" Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.665167 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555678-f2ldc" event={"ID":"6e31677c-4037-4cbc-9d34-5487431793e9","Type":"ContainerDied","Data":"45c44d1933bdfb1f99d54cecba14b4c68a52a137566e9c1b97d4dacb2c549a18"} Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.665819 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c44d1933bdfb1f99d54cecba14b4c68a52a137566e9c1b97d4dacb2c549a18" Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.665235 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555678-f2ldc" Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.909187 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:38:04 crc kubenswrapper[4926]: I0312 18:38:04.956780 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:38:05 crc kubenswrapper[4926]: I0312 18:38:05.051552 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555672-dq68h"] Mar 12 18:38:05 crc kubenswrapper[4926]: I0312 18:38:05.058516 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555672-dq68h"] Mar 12 18:38:05 crc kubenswrapper[4926]: I0312 18:38:05.141303 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ngnmz"] Mar 12 18:38:06 crc kubenswrapper[4926]: I0312 18:38:06.501759 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="609b8dbc-517d-4483-b02a-d7445cd2aa2f" path="/var/lib/kubelet/pods/609b8dbc-517d-4483-b02a-d7445cd2aa2f/volumes" Mar 12 18:38:06 crc kubenswrapper[4926]: I0312 18:38:06.680982 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ngnmz" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="registry-server" containerID="cri-o://74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f" gracePeriod=2 Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.119139 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.220204 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-catalog-content\") pod \"305b4cae-bc27-4c5d-8142-042446a753af\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.220330 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcbv8\" (UniqueName: \"kubernetes.io/projected/305b4cae-bc27-4c5d-8142-042446a753af-kube-api-access-mcbv8\") pod \"305b4cae-bc27-4c5d-8142-042446a753af\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.220471 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-utilities\") pod \"305b4cae-bc27-4c5d-8142-042446a753af\" (UID: \"305b4cae-bc27-4c5d-8142-042446a753af\") " Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.221146 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-utilities" (OuterVolumeSpecName: "utilities") pod "305b4cae-bc27-4c5d-8142-042446a753af" (UID: "305b4cae-bc27-4c5d-8142-042446a753af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.234805 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/305b4cae-bc27-4c5d-8142-042446a753af-kube-api-access-mcbv8" (OuterVolumeSpecName: "kube-api-access-mcbv8") pod "305b4cae-bc27-4c5d-8142-042446a753af" (UID: "305b4cae-bc27-4c5d-8142-042446a753af"). InnerVolumeSpecName "kube-api-access-mcbv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.322992 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcbv8\" (UniqueName: \"kubernetes.io/projected/305b4cae-bc27-4c5d-8142-042446a753af-kube-api-access-mcbv8\") on node \"crc\" DevicePath \"\"" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.323042 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.337995 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "305b4cae-bc27-4c5d-8142-042446a753af" (UID: "305b4cae-bc27-4c5d-8142-042446a753af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.424474 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/305b4cae-bc27-4c5d-8142-042446a753af-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.694660 4926 generic.go:334] "Generic (PLEG): container finished" podID="305b4cae-bc27-4c5d-8142-042446a753af" containerID="74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f" exitCode=0 Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.694762 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ngnmz" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.694789 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ngnmz" event={"ID":"305b4cae-bc27-4c5d-8142-042446a753af","Type":"ContainerDied","Data":"74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f"} Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.695294 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ngnmz" event={"ID":"305b4cae-bc27-4c5d-8142-042446a753af","Type":"ContainerDied","Data":"27c46b747fc3b972108b73d6610b46b16d2ab44f98be0213b6dafd0b1655a885"} Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.695328 4926 scope.go:117] "RemoveContainer" containerID="74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.738753 4926 scope.go:117] "RemoveContainer" containerID="45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.745634 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ngnmz"] Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.753941 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ngnmz"] Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.766414 4926 scope.go:117] "RemoveContainer" containerID="3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.810013 4926 scope.go:117] "RemoveContainer" containerID="74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f" Mar 12 18:38:07 crc kubenswrapper[4926]: E0312 18:38:07.810571 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f\": container with ID starting with 74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f not found: ID does not exist" containerID="74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.810605 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f"} err="failed to get container status \"74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f\": rpc error: code = NotFound desc = could not find container \"74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f\": container with ID starting with 74ce600f7c985ed38e0c138bcf27081c7bc8d0ed8d001ebf2fdb7df3b4033d0f not found: ID does not exist" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.810631 4926 scope.go:117] "RemoveContainer" containerID="45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3" Mar 12 18:38:07 crc kubenswrapper[4926]: E0312 18:38:07.810960 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3\": container with ID starting with 45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3 not found: ID does not exist" containerID="45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.810990 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3"} err="failed to get container status \"45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3\": rpc error: code = NotFound desc = could not find container \"45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3\": container with ID starting with 45de0ce8f8fe963be3a32d9a7c53fa9cb9f050130db08052d5f428486f761bb3 not found: ID does not exist" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.811010 4926 scope.go:117] "RemoveContainer" containerID="3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91" Mar 12 18:38:07 crc kubenswrapper[4926]: E0312 18:38:07.811466 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91\": container with ID starting with 3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91 not found: ID does not exist" containerID="3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91" Mar 12 18:38:07 crc kubenswrapper[4926]: I0312 18:38:07.811493 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91"} err="failed to get container status \"3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91\": rpc error: code = NotFound desc = could not find container \"3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91\": container with ID starting with 3db3234f981f4766633cb03a8859373ca4c860ed57313cf0bf6a95d3e1aa5e91 not found: ID does not exist" Mar 12 18:38:08 crc kubenswrapper[4926]: I0312 18:38:08.502308 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="305b4cae-bc27-4c5d-8142-042446a753af" path="/var/lib/kubelet/pods/305b4cae-bc27-4c5d-8142-042446a753af/volumes" Mar 12 18:38:09 crc kubenswrapper[4926]: I0312 18:38:09.849638 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-58dtw_16ea4f33-38fb-42c7-9c85-67c443f0b3a4/kube-rbac-proxy/0.log" Mar 12 18:38:09 crc kubenswrapper[4926]: I0312 18:38:09.887091 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-58dtw_16ea4f33-38fb-42c7-9c85-67c443f0b3a4/controller/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.102123 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-frr-files/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.295928 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-reloader/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.298334 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-reloader/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.321899 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-metrics/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.390859 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-frr-files/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.495786 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-metrics/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.500236 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-frr-files/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.544116 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-reloader/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.600779 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-metrics/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.710183 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-frr-files/0.log" Mar 12 18:38:10 crc kubenswrapper[4926]: I0312 18:38:10.777833 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-reloader/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.003876 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/controller/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.009495 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/cp-metrics/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.205040 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/frr-metrics/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.205353 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/kube-rbac-proxy/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.232467 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/kube-rbac-proxy-frr/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.451989 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/reloader/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.481116 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-j2n6d_036c2795-2942-4cc8-9a91-6cc48cbe7521/frr-k8s-webhook-server/0.log" Mar 12 18:38:11 crc kubenswrapper[4926]: I0312 18:38:11.801132 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7cf578c5b8-z4gn8_8a847a81-61ae-42e3-9866-c25b68fd77cb/manager/0.log" Mar 12 18:38:12 crc kubenswrapper[4926]: I0312 18:38:12.021683 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-79dbc878dc-w7f58_4d6f4326-2022-436b-9523-383aae3fd5cd/webhook-server/0.log" Mar 12 18:38:12 crc kubenswrapper[4926]: I0312 18:38:12.115945 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6bt9q_d10f0ca7-6fc4-4e6a-815c-ad5a1db16350/kube-rbac-proxy/0.log" Mar 12 18:38:12 crc kubenswrapper[4926]: I0312 18:38:12.338086 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9p72p_53da3fff-e3f4-4b9d-a887-f5a28f986107/frr/0.log" Mar 12 18:38:12 crc kubenswrapper[4926]: I0312 18:38:12.555346 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6bt9q_d10f0ca7-6fc4-4e6a-815c-ad5a1db16350/speaker/0.log" Mar 12 18:38:19 crc kubenswrapper[4926]: I0312 18:38:19.372007 4926 scope.go:117] "RemoveContainer" containerID="95bf3d49a01a19202e8eb63fbef45841e7cc791e3f3ccb3272c728981fc71d7d" Mar 12 18:38:25 crc kubenswrapper[4926]: I0312 18:38:25.409237 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq_062d1c31-cb0c-4470-bf74-0fb541319609/util/0.log" Mar 12 18:38:25 crc kubenswrapper[4926]: I0312 18:38:25.635898 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq_062d1c31-cb0c-4470-bf74-0fb541319609/util/0.log" Mar 12 18:38:25 crc kubenswrapper[4926]: I0312 18:38:25.646991 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq_062d1c31-cb0c-4470-bf74-0fb541319609/pull/0.log" Mar 12 18:38:25 crc kubenswrapper[4926]: I0312 18:38:25.647110 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq_062d1c31-cb0c-4470-bf74-0fb541319609/pull/0.log" Mar 12 18:38:25 crc kubenswrapper[4926]: I0312 18:38:25.819914 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq_062d1c31-cb0c-4470-bf74-0fb541319609/pull/0.log" Mar 12 18:38:25 crc kubenswrapper[4926]: I0312 18:38:25.856317 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq_062d1c31-cb0c-4470-bf74-0fb541319609/extract/0.log" Mar 12 18:38:25 crc kubenswrapper[4926]: I0312 18:38:25.865875 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874j5hnq_062d1c31-cb0c-4470-bf74-0fb541319609/util/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.001735 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg_a46be1f3-50fc-45b5-a480-98d9763db69d/util/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.149075 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg_a46be1f3-50fc-45b5-a480-98d9763db69d/util/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.157145 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg_a46be1f3-50fc-45b5-a480-98d9763db69d/pull/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.199143 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg_a46be1f3-50fc-45b5-a480-98d9763db69d/pull/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.397020 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg_a46be1f3-50fc-45b5-a480-98d9763db69d/extract/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.404248 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg_a46be1f3-50fc-45b5-a480-98d9763db69d/util/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.413074 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1pp5dg_a46be1f3-50fc-45b5-a480-98d9763db69d/pull/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.591076 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-t8lwn_0ef972e3-81c7-4b62-aa07-4939aef86a2d/extract-utilities/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.741106 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-t8lwn_0ef972e3-81c7-4b62-aa07-4939aef86a2d/extract-utilities/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.755966 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-t8lwn_0ef972e3-81c7-4b62-aa07-4939aef86a2d/extract-content/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.804292 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-t8lwn_0ef972e3-81c7-4b62-aa07-4939aef86a2d/extract-content/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.953100 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-t8lwn_0ef972e3-81c7-4b62-aa07-4939aef86a2d/extract-utilities/0.log" Mar 12 18:38:26 crc kubenswrapper[4926]: I0312 18:38:26.997124 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-t8lwn_0ef972e3-81c7-4b62-aa07-4939aef86a2d/extract-content/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.146068 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5k9gj_0817ed2e-acdd-41b8-b210-84281525839f/extract-utilities/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.365068 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-t8lwn_0ef972e3-81c7-4b62-aa07-4939aef86a2d/registry-server/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.422679 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5k9gj_0817ed2e-acdd-41b8-b210-84281525839f/extract-content/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.454327 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5k9gj_0817ed2e-acdd-41b8-b210-84281525839f/extract-utilities/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.476668 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5k9gj_0817ed2e-acdd-41b8-b210-84281525839f/extract-content/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.593529 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5k9gj_0817ed2e-acdd-41b8-b210-84281525839f/extract-utilities/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.639586 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5k9gj_0817ed2e-acdd-41b8-b210-84281525839f/extract-content/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.788196 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-5k9gj_0817ed2e-acdd-41b8-b210-84281525839f/registry-server/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.856714 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-d9gx5_daeebaaf-6a69-436e-b341-36fae756599e/marketplace-operator/0.log" Mar 12 18:38:27 crc kubenswrapper[4926]: I0312 18:38:27.954122 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-265z5_0f746e64-bce3-4f58-b789-0f5573e28847/extract-utilities/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.106361 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-265z5_0f746e64-bce3-4f58-b789-0f5573e28847/extract-utilities/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.135741 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-265z5_0f746e64-bce3-4f58-b789-0f5573e28847/extract-content/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.147061 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-265z5_0f746e64-bce3-4f58-b789-0f5573e28847/extract-content/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.351625 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-265z5_0f746e64-bce3-4f58-b789-0f5573e28847/extract-utilities/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.361498 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-265z5_0f746e64-bce3-4f58-b789-0f5573e28847/extract-content/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.417350 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-265z5_0f746e64-bce3-4f58-b789-0f5573e28847/registry-server/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.560737 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-672mz_c9f79c44-e93f-48ba-9f2d-8a5b61a86089/extract-utilities/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.689991 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-672mz_c9f79c44-e93f-48ba-9f2d-8a5b61a86089/extract-utilities/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.710518 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-672mz_c9f79c44-e93f-48ba-9f2d-8a5b61a86089/extract-content/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.737130 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-672mz_c9f79c44-e93f-48ba-9f2d-8a5b61a86089/extract-content/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.871735 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-672mz_c9f79c44-e93f-48ba-9f2d-8a5b61a86089/extract-utilities/0.log" Mar 12 18:38:28 crc kubenswrapper[4926]: I0312 18:38:28.908057 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-672mz_c9f79c44-e93f-48ba-9f2d-8a5b61a86089/extract-content/0.log" Mar 12 18:38:29 crc kubenswrapper[4926]: I0312 18:38:29.234911 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-672mz_c9f79c44-e93f-48ba-9f2d-8a5b61a86089/registry-server/0.log" Mar 12 18:39:26 crc kubenswrapper[4926]: I0312 18:39:26.817588 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:39:26 crc kubenswrapper[4926]: I0312 18:39:26.818096 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:39:56 crc kubenswrapper[4926]: I0312 18:39:56.818129 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:39:56 crc kubenswrapper[4926]: I0312 18:39:56.820352 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.153630 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555680-cfpm6"] Mar 12 18:40:00 crc kubenswrapper[4926]: E0312 18:40:00.154649 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="registry-server" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.154665 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="registry-server" Mar 12 18:40:00 crc kubenswrapper[4926]: E0312 18:40:00.154882 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e31677c-4037-4cbc-9d34-5487431793e9" containerName="oc" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.154890 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e31677c-4037-4cbc-9d34-5487431793e9" containerName="oc" Mar 12 18:40:00 crc kubenswrapper[4926]: E0312 18:40:00.154915 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="extract-content" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.154923 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="extract-content" Mar 12 18:40:00 crc kubenswrapper[4926]: E0312 18:40:00.154941 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="extract-utilities" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.154949 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="extract-utilities" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.155136 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e31677c-4037-4cbc-9d34-5487431793e9" containerName="oc" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.155159 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="305b4cae-bc27-4c5d-8142-042446a753af" containerName="registry-server" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.156007 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555680-cfpm6" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.161073 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.161363 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.161594 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.166092 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555680-cfpm6"] Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.218190 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m722\" (UniqueName: \"kubernetes.io/projected/d3cd40c1-ffde-4c68-b932-d84c58ec4364-kube-api-access-2m722\") pod \"auto-csr-approver-29555680-cfpm6\" (UID: \"d3cd40c1-ffde-4c68-b932-d84c58ec4364\") " pod="openshift-infra/auto-csr-approver-29555680-cfpm6" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.320939 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m722\" (UniqueName: \"kubernetes.io/projected/d3cd40c1-ffde-4c68-b932-d84c58ec4364-kube-api-access-2m722\") pod \"auto-csr-approver-29555680-cfpm6\" (UID: \"d3cd40c1-ffde-4c68-b932-d84c58ec4364\") " pod="openshift-infra/auto-csr-approver-29555680-cfpm6" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.348758 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m722\" (UniqueName: \"kubernetes.io/projected/d3cd40c1-ffde-4c68-b932-d84c58ec4364-kube-api-access-2m722\") pod \"auto-csr-approver-29555680-cfpm6\" (UID: \"d3cd40c1-ffde-4c68-b932-d84c58ec4364\") " pod="openshift-infra/auto-csr-approver-29555680-cfpm6" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.475470 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555680-cfpm6" Mar 12 18:40:00 crc kubenswrapper[4926]: I0312 18:40:00.985186 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555680-cfpm6"] Mar 12 18:40:01 crc kubenswrapper[4926]: I0312 18:40:01.790258 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555680-cfpm6" event={"ID":"d3cd40c1-ffde-4c68-b932-d84c58ec4364","Type":"ContainerStarted","Data":"300c905d81cac499e9f45131e33911c3e5941a151d34623a2565c4b79bcdc657"} Mar 12 18:40:03 crc kubenswrapper[4926]: I0312 18:40:03.811921 4926 generic.go:334] "Generic (PLEG): container finished" podID="d3cd40c1-ffde-4c68-b932-d84c58ec4364" containerID="ed5e9b74d75807b6ccd7ef1a2735ad64a65a226a6e6c26e9d80b7971ca9a6cd3" exitCode=0 Mar 12 18:40:03 crc kubenswrapper[4926]: I0312 18:40:03.812199 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555680-cfpm6" event={"ID":"d3cd40c1-ffde-4c68-b932-d84c58ec4364","Type":"ContainerDied","Data":"ed5e9b74d75807b6ccd7ef1a2735ad64a65a226a6e6c26e9d80b7971ca9a6cd3"} Mar 12 18:40:05 crc kubenswrapper[4926]: I0312 18:40:05.243048 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555680-cfpm6" Mar 12 18:40:05 crc kubenswrapper[4926]: I0312 18:40:05.337593 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m722\" (UniqueName: \"kubernetes.io/projected/d3cd40c1-ffde-4c68-b932-d84c58ec4364-kube-api-access-2m722\") pod \"d3cd40c1-ffde-4c68-b932-d84c58ec4364\" (UID: \"d3cd40c1-ffde-4c68-b932-d84c58ec4364\") " Mar 12 18:40:05 crc kubenswrapper[4926]: I0312 18:40:05.343728 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3cd40c1-ffde-4c68-b932-d84c58ec4364-kube-api-access-2m722" (OuterVolumeSpecName: "kube-api-access-2m722") pod "d3cd40c1-ffde-4c68-b932-d84c58ec4364" (UID: "d3cd40c1-ffde-4c68-b932-d84c58ec4364"). InnerVolumeSpecName "kube-api-access-2m722". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:40:05 crc kubenswrapper[4926]: I0312 18:40:05.441293 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m722\" (UniqueName: \"kubernetes.io/projected/d3cd40c1-ffde-4c68-b932-d84c58ec4364-kube-api-access-2m722\") on node \"crc\" DevicePath \"\"" Mar 12 18:40:05 crc kubenswrapper[4926]: I0312 18:40:05.835497 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555680-cfpm6" event={"ID":"d3cd40c1-ffde-4c68-b932-d84c58ec4364","Type":"ContainerDied","Data":"300c905d81cac499e9f45131e33911c3e5941a151d34623a2565c4b79bcdc657"} Mar 12 18:40:05 crc kubenswrapper[4926]: I0312 18:40:05.835569 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="300c905d81cac499e9f45131e33911c3e5941a151d34623a2565c4b79bcdc657" Mar 12 18:40:05 crc kubenswrapper[4926]: I0312 18:40:05.835627 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555680-cfpm6" Mar 12 18:40:06 crc kubenswrapper[4926]: I0312 18:40:06.334944 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555674-2b7bv"] Mar 12 18:40:06 crc kubenswrapper[4926]: I0312 18:40:06.342651 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555674-2b7bv"] Mar 12 18:40:06 crc kubenswrapper[4926]: I0312 18:40:06.507906 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="540a72de-64b2-41af-9caf-95895f11cb79" path="/var/lib/kubelet/pods/540a72de-64b2-41af-9caf-95895f11cb79/volumes" Mar 12 18:40:11 crc kubenswrapper[4926]: I0312 18:40:11.906584 4926 generic.go:334] "Generic (PLEG): container finished" podID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerID="4dbea05e4f0fb083c128a04cb1592adbe2b7259461b2a690428955b1ed545c5d" exitCode=0 Mar 12 18:40:11 crc kubenswrapper[4926]: I0312 18:40:11.906728 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" event={"ID":"2562de93-d8c3-4055-8ab2-3c55b4f3c830","Type":"ContainerDied","Data":"4dbea05e4f0fb083c128a04cb1592adbe2b7259461b2a690428955b1ed545c5d"} Mar 12 18:40:11 crc kubenswrapper[4926]: I0312 18:40:11.908620 4926 scope.go:117] "RemoveContainer" containerID="4dbea05e4f0fb083c128a04cb1592adbe2b7259461b2a690428955b1ed545c5d" Mar 12 18:40:12 crc kubenswrapper[4926]: I0312 18:40:12.068678 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nz8z6_must-gather-x4gfw_2562de93-d8c3-4055-8ab2-3c55b4f3c830/gather/0.log" Mar 12 18:40:19 crc kubenswrapper[4926]: I0312 18:40:19.495482 4926 scope.go:117] "RemoveContainer" containerID="804c89c001a28f1eae59ba1c1cb327e0cc2c54749f5a07cd9cd07164cc1b86c4" Mar 12 18:40:19 crc kubenswrapper[4926]: I0312 18:40:19.677489 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nz8z6/must-gather-x4gfw"] Mar 12 18:40:19 crc kubenswrapper[4926]: I0312 18:40:19.678075 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerName="copy" containerID="cri-o://9983eed33b5073c82867ffe643c25901fb02264d34aa9779308efc3e4497884a" gracePeriod=2 Mar 12 18:40:19 crc kubenswrapper[4926]: I0312 18:40:19.687102 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nz8z6/must-gather-x4gfw"] Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.003088 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nz8z6_must-gather-x4gfw_2562de93-d8c3-4055-8ab2-3c55b4f3c830/copy/0.log" Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.003552 4926 generic.go:334] "Generic (PLEG): container finished" podID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerID="9983eed33b5073c82867ffe643c25901fb02264d34aa9779308efc3e4497884a" exitCode=143 Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.210251 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nz8z6_must-gather-x4gfw_2562de93-d8c3-4055-8ab2-3c55b4f3c830/copy/0.log" Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.211005 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.299117 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w44tw\" (UniqueName: \"kubernetes.io/projected/2562de93-d8c3-4055-8ab2-3c55b4f3c830-kube-api-access-w44tw\") pod \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.299232 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2562de93-d8c3-4055-8ab2-3c55b4f3c830-must-gather-output\") pod \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\" (UID: \"2562de93-d8c3-4055-8ab2-3c55b4f3c830\") " Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.310735 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2562de93-d8c3-4055-8ab2-3c55b4f3c830-kube-api-access-w44tw" (OuterVolumeSpecName: "kube-api-access-w44tw") pod "2562de93-d8c3-4055-8ab2-3c55b4f3c830" (UID: "2562de93-d8c3-4055-8ab2-3c55b4f3c830"). InnerVolumeSpecName "kube-api-access-w44tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.401619 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w44tw\" (UniqueName: \"kubernetes.io/projected/2562de93-d8c3-4055-8ab2-3c55b4f3c830-kube-api-access-w44tw\") on node \"crc\" DevicePath \"\"" Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.458612 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2562de93-d8c3-4055-8ab2-3c55b4f3c830-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2562de93-d8c3-4055-8ab2-3c55b4f3c830" (UID: "2562de93-d8c3-4055-8ab2-3c55b4f3c830"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.500733 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" path="/var/lib/kubelet/pods/2562de93-d8c3-4055-8ab2-3c55b4f3c830/volumes" Mar 12 18:40:20 crc kubenswrapper[4926]: I0312 18:40:20.502573 4926 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2562de93-d8c3-4055-8ab2-3c55b4f3c830-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 12 18:40:21 crc kubenswrapper[4926]: I0312 18:40:21.011905 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nz8z6_must-gather-x4gfw_2562de93-d8c3-4055-8ab2-3c55b4f3c830/copy/0.log" Mar 12 18:40:21 crc kubenswrapper[4926]: I0312 18:40:21.012215 4926 scope.go:117] "RemoveContainer" containerID="9983eed33b5073c82867ffe643c25901fb02264d34aa9779308efc3e4497884a" Mar 12 18:40:21 crc kubenswrapper[4926]: I0312 18:40:21.012278 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nz8z6/must-gather-x4gfw" Mar 12 18:40:21 crc kubenswrapper[4926]: I0312 18:40:21.031483 4926 scope.go:117] "RemoveContainer" containerID="4dbea05e4f0fb083c128a04cb1592adbe2b7259461b2a690428955b1ed545c5d" Mar 12 18:40:26 crc kubenswrapper[4926]: I0312 18:40:26.817593 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:40:26 crc kubenswrapper[4926]: I0312 18:40:26.818183 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:40:26 crc kubenswrapper[4926]: I0312 18:40:26.818248 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:40:26 crc kubenswrapper[4926]: I0312 18:40:26.819258 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2201cfc89c392b2d9df343a46350dae7f7620675e753c895f182c00c5eb9467c"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:40:26 crc kubenswrapper[4926]: I0312 18:40:26.819323 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://2201cfc89c392b2d9df343a46350dae7f7620675e753c895f182c00c5eb9467c" gracePeriod=600 Mar 12 18:40:27 crc kubenswrapper[4926]: I0312 18:40:27.090710 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="2201cfc89c392b2d9df343a46350dae7f7620675e753c895f182c00c5eb9467c" exitCode=0 Mar 12 18:40:27 crc kubenswrapper[4926]: I0312 18:40:27.091098 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"2201cfc89c392b2d9df343a46350dae7f7620675e753c895f182c00c5eb9467c"} Mar 12 18:40:27 crc kubenswrapper[4926]: I0312 18:40:27.091146 4926 scope.go:117] "RemoveContainer" containerID="842a75b054aae388d59ac83e483a69d941997a23cf47d9012a53fea65a005b5e" Mar 12 18:40:28 crc kubenswrapper[4926]: I0312 18:40:28.099924 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerStarted","Data":"79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2"} Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.976616 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4fx77"] Mar 12 18:40:58 crc kubenswrapper[4926]: E0312 18:40:58.984043 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerName="copy" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.984072 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerName="copy" Mar 12 18:40:58 crc kubenswrapper[4926]: E0312 18:40:58.984092 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3cd40c1-ffde-4c68-b932-d84c58ec4364" containerName="oc" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.984102 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3cd40c1-ffde-4c68-b932-d84c58ec4364" containerName="oc" Mar 12 18:40:58 crc kubenswrapper[4926]: E0312 18:40:58.984118 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerName="gather" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.984125 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerName="gather" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.984333 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerName="copy" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.984348 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2562de93-d8c3-4055-8ab2-3c55b4f3c830" containerName="gather" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.984358 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3cd40c1-ffde-4c68-b932-d84c58ec4364" containerName="oc" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.986072 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:58 crc kubenswrapper[4926]: I0312 18:40:58.998502 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4fx77"] Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.036626 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-utilities\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.037082 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbtsw\" (UniqueName: \"kubernetes.io/projected/d7e47942-3557-4107-a493-c18e4d174d1e-kube-api-access-bbtsw\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.037270 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-catalog-content\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.139610 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-utilities\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.139767 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbtsw\" (UniqueName: \"kubernetes.io/projected/d7e47942-3557-4107-a493-c18e4d174d1e-kube-api-access-bbtsw\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.139840 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-catalog-content\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.140181 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-utilities\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.140255 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-catalog-content\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.165532 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbtsw\" (UniqueName: \"kubernetes.io/projected/d7e47942-3557-4107-a493-c18e4d174d1e-kube-api-access-bbtsw\") pod \"certified-operators-4fx77\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.311651 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:40:59 crc kubenswrapper[4926]: I0312 18:40:59.786780 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4fx77"] Mar 12 18:41:00 crc kubenswrapper[4926]: I0312 18:41:00.424319 4926 generic.go:334] "Generic (PLEG): container finished" podID="d7e47942-3557-4107-a493-c18e4d174d1e" containerID="5571f3a9f466caaa9333a8b56f85c61cee8d3b7a08a46b6c991362072f21c04e" exitCode=0 Mar 12 18:41:00 crc kubenswrapper[4926]: I0312 18:41:00.424383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fx77" event={"ID":"d7e47942-3557-4107-a493-c18e4d174d1e","Type":"ContainerDied","Data":"5571f3a9f466caaa9333a8b56f85c61cee8d3b7a08a46b6c991362072f21c04e"} Mar 12 18:41:00 crc kubenswrapper[4926]: I0312 18:41:00.424719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fx77" event={"ID":"d7e47942-3557-4107-a493-c18e4d174d1e","Type":"ContainerStarted","Data":"81fcc693eefa118dd047e3cba7618229952a64d7026c69e5f1def03ed6e8bd7d"} Mar 12 18:41:01 crc kubenswrapper[4926]: I0312 18:41:01.435728 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fx77" event={"ID":"d7e47942-3557-4107-a493-c18e4d174d1e","Type":"ContainerStarted","Data":"a16e7dd57edf66ac5ce9d3d0f07997272a6bc5cd63023c443aa9cd54a9bf9d1d"} Mar 12 18:41:02 crc kubenswrapper[4926]: I0312 18:41:02.447113 4926 generic.go:334] "Generic (PLEG): container finished" podID="d7e47942-3557-4107-a493-c18e4d174d1e" containerID="a16e7dd57edf66ac5ce9d3d0f07997272a6bc5cd63023c443aa9cd54a9bf9d1d" exitCode=0 Mar 12 18:41:02 crc kubenswrapper[4926]: I0312 18:41:02.447192 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fx77" event={"ID":"d7e47942-3557-4107-a493-c18e4d174d1e","Type":"ContainerDied","Data":"a16e7dd57edf66ac5ce9d3d0f07997272a6bc5cd63023c443aa9cd54a9bf9d1d"} Mar 12 18:41:04 crc kubenswrapper[4926]: I0312 18:41:04.474874 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fx77" event={"ID":"d7e47942-3557-4107-a493-c18e4d174d1e","Type":"ContainerStarted","Data":"f9735ccc51d66ef1ac0075c04afb2fe41d5c57065507fc42e109f6d09e13bd75"} Mar 12 18:41:04 crc kubenswrapper[4926]: I0312 18:41:04.508722 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4fx77" podStartSLOduration=3.469928134 podStartE2EDuration="6.508704734s" podCreationTimestamp="2026-03-12 18:40:58 +0000 UTC" firstStartedPulling="2026-03-12 18:41:00.426934449 +0000 UTC m=+2300.795560812" lastFinishedPulling="2026-03-12 18:41:03.465711069 +0000 UTC m=+2303.834337412" observedRunningTime="2026-03-12 18:41:04.502669952 +0000 UTC m=+2304.871296295" watchObservedRunningTime="2026-03-12 18:41:04.508704734 +0000 UTC m=+2304.877331077" Mar 12 18:41:09 crc kubenswrapper[4926]: I0312 18:41:09.311800 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:41:09 crc kubenswrapper[4926]: I0312 18:41:09.312325 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:41:09 crc kubenswrapper[4926]: I0312 18:41:09.363061 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:41:09 crc kubenswrapper[4926]: I0312 18:41:09.579792 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:41:09 crc kubenswrapper[4926]: I0312 18:41:09.624253 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4fx77"] Mar 12 18:41:11 crc kubenswrapper[4926]: I0312 18:41:11.553332 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4fx77" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="registry-server" containerID="cri-o://f9735ccc51d66ef1ac0075c04afb2fe41d5c57065507fc42e109f6d09e13bd75" gracePeriod=2 Mar 12 18:41:12 crc kubenswrapper[4926]: I0312 18:41:12.567469 4926 generic.go:334] "Generic (PLEG): container finished" podID="d7e47942-3557-4107-a493-c18e4d174d1e" containerID="f9735ccc51d66ef1ac0075c04afb2fe41d5c57065507fc42e109f6d09e13bd75" exitCode=0 Mar 12 18:41:12 crc kubenswrapper[4926]: I0312 18:41:12.567524 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fx77" event={"ID":"d7e47942-3557-4107-a493-c18e4d174d1e","Type":"ContainerDied","Data":"f9735ccc51d66ef1ac0075c04afb2fe41d5c57065507fc42e109f6d09e13bd75"} Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.393508 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.530567 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbtsw\" (UniqueName: \"kubernetes.io/projected/d7e47942-3557-4107-a493-c18e4d174d1e-kube-api-access-bbtsw\") pod \"d7e47942-3557-4107-a493-c18e4d174d1e\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.530665 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-utilities\") pod \"d7e47942-3557-4107-a493-c18e4d174d1e\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.530734 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-catalog-content\") pod \"d7e47942-3557-4107-a493-c18e4d174d1e\" (UID: \"d7e47942-3557-4107-a493-c18e4d174d1e\") " Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.532855 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-utilities" (OuterVolumeSpecName: "utilities") pod "d7e47942-3557-4107-a493-c18e4d174d1e" (UID: "d7e47942-3557-4107-a493-c18e4d174d1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.540192 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e47942-3557-4107-a493-c18e4d174d1e-kube-api-access-bbtsw" (OuterVolumeSpecName: "kube-api-access-bbtsw") pod "d7e47942-3557-4107-a493-c18e4d174d1e" (UID: "d7e47942-3557-4107-a493-c18e4d174d1e"). InnerVolumeSpecName "kube-api-access-bbtsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.583669 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4fx77" event={"ID":"d7e47942-3557-4107-a493-c18e4d174d1e","Type":"ContainerDied","Data":"81fcc693eefa118dd047e3cba7618229952a64d7026c69e5f1def03ed6e8bd7d"} Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.583731 4926 scope.go:117] "RemoveContainer" containerID="f9735ccc51d66ef1ac0075c04afb2fe41d5c57065507fc42e109f6d09e13bd75" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.583731 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4fx77" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.595175 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7e47942-3557-4107-a493-c18e4d174d1e" (UID: "d7e47942-3557-4107-a493-c18e4d174d1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.632785 4926 scope.go:117] "RemoveContainer" containerID="a16e7dd57edf66ac5ce9d3d0f07997272a6bc5cd63023c443aa9cd54a9bf9d1d" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.633370 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.633405 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7e47942-3557-4107-a493-c18e4d174d1e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.633420 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbtsw\" (UniqueName: \"kubernetes.io/projected/d7e47942-3557-4107-a493-c18e4d174d1e-kube-api-access-bbtsw\") on node \"crc\" DevicePath \"\"" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.661380 4926 scope.go:117] "RemoveContainer" containerID="5571f3a9f466caaa9333a8b56f85c61cee8d3b7a08a46b6c991362072f21c04e" Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.929022 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4fx77"] Mar 12 18:41:13 crc kubenswrapper[4926]: I0312 18:41:13.939372 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4fx77"] Mar 12 18:41:14 crc kubenswrapper[4926]: I0312 18:41:14.500595 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" path="/var/lib/kubelet/pods/d7e47942-3557-4107-a493-c18e4d174d1e/volumes" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.142411 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-822np"] Mar 12 18:41:20 crc kubenswrapper[4926]: E0312 18:41:20.143352 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="extract-utilities" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.143369 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="extract-utilities" Mar 12 18:41:20 crc kubenswrapper[4926]: E0312 18:41:20.143390 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="registry-server" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.143399 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="registry-server" Mar 12 18:41:20 crc kubenswrapper[4926]: E0312 18:41:20.143418 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="extract-content" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.143429 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="extract-content" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.143706 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e47942-3557-4107-a493-c18e4d174d1e" containerName="registry-server" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.145299 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.161800 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-822np"] Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.265707 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-utilities\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.266309 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgm8j\" (UniqueName: \"kubernetes.io/projected/bd48d2ff-4171-4a1a-a3f7-d117afc68847-kube-api-access-zgm8j\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.266351 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-catalog-content\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.368575 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgm8j\" (UniqueName: \"kubernetes.io/projected/bd48d2ff-4171-4a1a-a3f7-d117afc68847-kube-api-access-zgm8j\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.368633 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-catalog-content\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.368768 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-utilities\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.369193 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-utilities\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.369401 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-catalog-content\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.395960 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgm8j\" (UniqueName: \"kubernetes.io/projected/bd48d2ff-4171-4a1a-a3f7-d117afc68847-kube-api-access-zgm8j\") pod \"redhat-marketplace-822np\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.479303 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:20 crc kubenswrapper[4926]: I0312 18:41:20.923523 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-822np"] Mar 12 18:41:21 crc kubenswrapper[4926]: I0312 18:41:21.666189 4926 generic.go:334] "Generic (PLEG): container finished" podID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerID="254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab" exitCode=0 Mar 12 18:41:21 crc kubenswrapper[4926]: I0312 18:41:21.666513 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-822np" event={"ID":"bd48d2ff-4171-4a1a-a3f7-d117afc68847","Type":"ContainerDied","Data":"254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab"} Mar 12 18:41:21 crc kubenswrapper[4926]: I0312 18:41:21.666609 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-822np" event={"ID":"bd48d2ff-4171-4a1a-a3f7-d117afc68847","Type":"ContainerStarted","Data":"be102d0aa90eeef589f542e7c63da5e76a87f4a8ad4b9fbaa525d5b4bde3a687"} Mar 12 18:41:22 crc kubenswrapper[4926]: I0312 18:41:22.688468 4926 generic.go:334] "Generic (PLEG): container finished" podID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerID="987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0" exitCode=0 Mar 12 18:41:22 crc kubenswrapper[4926]: I0312 18:41:22.688581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-822np" event={"ID":"bd48d2ff-4171-4a1a-a3f7-d117afc68847","Type":"ContainerDied","Data":"987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0"} Mar 12 18:41:23 crc kubenswrapper[4926]: I0312 18:41:23.702299 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-822np" event={"ID":"bd48d2ff-4171-4a1a-a3f7-d117afc68847","Type":"ContainerStarted","Data":"9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea"} Mar 12 18:41:23 crc kubenswrapper[4926]: I0312 18:41:23.740404 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-822np" podStartSLOduration=2.277799989 podStartE2EDuration="3.740383244s" podCreationTimestamp="2026-03-12 18:41:20 +0000 UTC" firstStartedPulling="2026-03-12 18:41:21.672584631 +0000 UTC m=+2322.041211004" lastFinishedPulling="2026-03-12 18:41:23.135167916 +0000 UTC m=+2323.503794259" observedRunningTime="2026-03-12 18:41:23.734658262 +0000 UTC m=+2324.103284645" watchObservedRunningTime="2026-03-12 18:41:23.740383244 +0000 UTC m=+2324.109009587" Mar 12 18:41:30 crc kubenswrapper[4926]: I0312 18:41:30.480314 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:30 crc kubenswrapper[4926]: I0312 18:41:30.480920 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:30 crc kubenswrapper[4926]: I0312 18:41:30.537324 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:30 crc kubenswrapper[4926]: I0312 18:41:30.823812 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:30 crc kubenswrapper[4926]: I0312 18:41:30.880330 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-822np"] Mar 12 18:41:32 crc kubenswrapper[4926]: I0312 18:41:32.782028 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-822np" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="registry-server" containerID="cri-o://9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea" gracePeriod=2 Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.306949 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.442754 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgm8j\" (UniqueName: \"kubernetes.io/projected/bd48d2ff-4171-4a1a-a3f7-d117afc68847-kube-api-access-zgm8j\") pod \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.442876 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-catalog-content\") pod \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.442907 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-utilities\") pod \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\" (UID: \"bd48d2ff-4171-4a1a-a3f7-d117afc68847\") " Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.444655 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-utilities" (OuterVolumeSpecName: "utilities") pod "bd48d2ff-4171-4a1a-a3f7-d117afc68847" (UID: "bd48d2ff-4171-4a1a-a3f7-d117afc68847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.468362 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd48d2ff-4171-4a1a-a3f7-d117afc68847-kube-api-access-zgm8j" (OuterVolumeSpecName: "kube-api-access-zgm8j") pod "bd48d2ff-4171-4a1a-a3f7-d117afc68847" (UID: "bd48d2ff-4171-4a1a-a3f7-d117afc68847"). InnerVolumeSpecName "kube-api-access-zgm8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.485240 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd48d2ff-4171-4a1a-a3f7-d117afc68847" (UID: "bd48d2ff-4171-4a1a-a3f7-d117afc68847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.545499 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgm8j\" (UniqueName: \"kubernetes.io/projected/bd48d2ff-4171-4a1a-a3f7-d117afc68847-kube-api-access-zgm8j\") on node \"crc\" DevicePath \"\"" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.545529 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.545539 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd48d2ff-4171-4a1a-a3f7-d117afc68847-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.791632 4926 generic.go:334] "Generic (PLEG): container finished" podID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerID="9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea" exitCode=0 Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.791673 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-822np" event={"ID":"bd48d2ff-4171-4a1a-a3f7-d117afc68847","Type":"ContainerDied","Data":"9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea"} Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.791704 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-822np" event={"ID":"bd48d2ff-4171-4a1a-a3f7-d117afc68847","Type":"ContainerDied","Data":"be102d0aa90eeef589f542e7c63da5e76a87f4a8ad4b9fbaa525d5b4bde3a687"} Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.791720 4926 scope.go:117] "RemoveContainer" containerID="9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.791841 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-822np" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.816053 4926 scope.go:117] "RemoveContainer" containerID="987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.830682 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-822np"] Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.837999 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-822np"] Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.849288 4926 scope.go:117] "RemoveContainer" containerID="254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.905020 4926 scope.go:117] "RemoveContainer" containerID="9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea" Mar 12 18:41:33 crc kubenswrapper[4926]: E0312 18:41:33.905576 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea\": container with ID starting with 9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea not found: ID does not exist" containerID="9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.905630 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea"} err="failed to get container status \"9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea\": rpc error: code = NotFound desc = could not find container \"9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea\": container with ID starting with 9f46a3f2958fc6e08484fb487adaaf277fd41172828d236bc496aa7dc29cf0ea not found: ID does not exist" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.905665 4926 scope.go:117] "RemoveContainer" containerID="987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0" Mar 12 18:41:33 crc kubenswrapper[4926]: E0312 18:41:33.905969 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0\": container with ID starting with 987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0 not found: ID does not exist" containerID="987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.905990 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0"} err="failed to get container status \"987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0\": rpc error: code = NotFound desc = could not find container \"987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0\": container with ID starting with 987d4bae5548023fd3225ac0b17dd33eced3dc2e32196319e0e0ab5cbc06f4b0 not found: ID does not exist" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.906004 4926 scope.go:117] "RemoveContainer" containerID="254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab" Mar 12 18:41:33 crc kubenswrapper[4926]: E0312 18:41:33.906267 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab\": container with ID starting with 254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab not found: ID does not exist" containerID="254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab" Mar 12 18:41:33 crc kubenswrapper[4926]: I0312 18:41:33.906292 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab"} err="failed to get container status \"254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab\": rpc error: code = NotFound desc = could not find container \"254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab\": container with ID starting with 254ef0ca2b3871aa0700b6c4457562c54faac494d6dcd4ad0aa5d755780cd7ab not found: ID does not exist" Mar 12 18:41:34 crc kubenswrapper[4926]: I0312 18:41:34.504524 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" path="/var/lib/kubelet/pods/bd48d2ff-4171-4a1a-a3f7-d117afc68847/volumes" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.154201 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555682-rsgxm"] Mar 12 18:42:00 crc kubenswrapper[4926]: E0312 18:42:00.155150 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="extract-content" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.155165 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="extract-content" Mar 12 18:42:00 crc kubenswrapper[4926]: E0312 18:42:00.155193 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="registry-server" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.155201 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="registry-server" Mar 12 18:42:00 crc kubenswrapper[4926]: E0312 18:42:00.155223 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="extract-utilities" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.155233 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="extract-utilities" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.155460 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd48d2ff-4171-4a1a-a3f7-d117afc68847" containerName="registry-server" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.164867 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555682-rsgxm" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.168524 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.169063 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.169215 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.177185 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555682-rsgxm"] Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.192836 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x268q\" (UniqueName: \"kubernetes.io/projected/2955f1b9-7c06-4681-b024-65ce46309648-kube-api-access-x268q\") pod \"auto-csr-approver-29555682-rsgxm\" (UID: \"2955f1b9-7c06-4681-b024-65ce46309648\") " pod="openshift-infra/auto-csr-approver-29555682-rsgxm" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.294546 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x268q\" (UniqueName: \"kubernetes.io/projected/2955f1b9-7c06-4681-b024-65ce46309648-kube-api-access-x268q\") pod \"auto-csr-approver-29555682-rsgxm\" (UID: \"2955f1b9-7c06-4681-b024-65ce46309648\") " pod="openshift-infra/auto-csr-approver-29555682-rsgxm" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.314386 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x268q\" (UniqueName: \"kubernetes.io/projected/2955f1b9-7c06-4681-b024-65ce46309648-kube-api-access-x268q\") pod \"auto-csr-approver-29555682-rsgxm\" (UID: \"2955f1b9-7c06-4681-b024-65ce46309648\") " pod="openshift-infra/auto-csr-approver-29555682-rsgxm" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.491096 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555682-rsgxm" Mar 12 18:42:00 crc kubenswrapper[4926]: I0312 18:42:00.942264 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555682-rsgxm"] Mar 12 18:42:00 crc kubenswrapper[4926]: W0312 18:42:00.948398 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2955f1b9_7c06_4681_b024_65ce46309648.slice/crio-b2956325606c98cf949c53ea15ba2eef378c1602cb590b8b636988f46dd4d2e9 WatchSource:0}: Error finding container b2956325606c98cf949c53ea15ba2eef378c1602cb590b8b636988f46dd4d2e9: Status 404 returned error can't find the container with id b2956325606c98cf949c53ea15ba2eef378c1602cb590b8b636988f46dd4d2e9 Mar 12 18:42:01 crc kubenswrapper[4926]: I0312 18:42:01.100763 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555682-rsgxm" event={"ID":"2955f1b9-7c06-4681-b024-65ce46309648","Type":"ContainerStarted","Data":"b2956325606c98cf949c53ea15ba2eef378c1602cb590b8b636988f46dd4d2e9"} Mar 12 18:42:03 crc kubenswrapper[4926]: I0312 18:42:03.119854 4926 generic.go:334] "Generic (PLEG): container finished" podID="2955f1b9-7c06-4681-b024-65ce46309648" containerID="ae4090da1d7102f7135d176cea5cebb506f68ec5d1e9004e09e2c6fdc15528d7" exitCode=0 Mar 12 18:42:03 crc kubenswrapper[4926]: I0312 18:42:03.120039 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555682-rsgxm" event={"ID":"2955f1b9-7c06-4681-b024-65ce46309648","Type":"ContainerDied","Data":"ae4090da1d7102f7135d176cea5cebb506f68ec5d1e9004e09e2c6fdc15528d7"} Mar 12 18:42:04 crc kubenswrapper[4926]: I0312 18:42:04.520212 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555682-rsgxm" Mar 12 18:42:04 crc kubenswrapper[4926]: I0312 18:42:04.696170 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x268q\" (UniqueName: \"kubernetes.io/projected/2955f1b9-7c06-4681-b024-65ce46309648-kube-api-access-x268q\") pod \"2955f1b9-7c06-4681-b024-65ce46309648\" (UID: \"2955f1b9-7c06-4681-b024-65ce46309648\") " Mar 12 18:42:04 crc kubenswrapper[4926]: I0312 18:42:04.701832 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2955f1b9-7c06-4681-b024-65ce46309648-kube-api-access-x268q" (OuterVolumeSpecName: "kube-api-access-x268q") pod "2955f1b9-7c06-4681-b024-65ce46309648" (UID: "2955f1b9-7c06-4681-b024-65ce46309648"). InnerVolumeSpecName "kube-api-access-x268q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:42:04 crc kubenswrapper[4926]: I0312 18:42:04.798752 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x268q\" (UniqueName: \"kubernetes.io/projected/2955f1b9-7c06-4681-b024-65ce46309648-kube-api-access-x268q\") on node \"crc\" DevicePath \"\"" Mar 12 18:42:05 crc kubenswrapper[4926]: I0312 18:42:05.145234 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555682-rsgxm" event={"ID":"2955f1b9-7c06-4681-b024-65ce46309648","Type":"ContainerDied","Data":"b2956325606c98cf949c53ea15ba2eef378c1602cb590b8b636988f46dd4d2e9"} Mar 12 18:42:05 crc kubenswrapper[4926]: I0312 18:42:05.145285 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2956325606c98cf949c53ea15ba2eef378c1602cb590b8b636988f46dd4d2e9" Mar 12 18:42:05 crc kubenswrapper[4926]: I0312 18:42:05.145347 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555682-rsgxm" Mar 12 18:42:05 crc kubenswrapper[4926]: I0312 18:42:05.603738 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555676-5d5p5"] Mar 12 18:42:05 crc kubenswrapper[4926]: I0312 18:42:05.612035 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555676-5d5p5"] Mar 12 18:42:06 crc kubenswrapper[4926]: I0312 18:42:06.509624 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1b045ea-1fff-43b3-9a66-48dd361e9f33" path="/var/lib/kubelet/pods/f1b045ea-1fff-43b3-9a66-48dd361e9f33/volumes" Mar 12 18:42:19 crc kubenswrapper[4926]: I0312 18:42:19.688077 4926 scope.go:117] "RemoveContainer" containerID="cfe5fa51265434fa1414bed949d3e915d15d11d193d0cf363993fcfab40b91d4" Mar 12 18:42:56 crc kubenswrapper[4926]: I0312 18:42:56.818186 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:42:56 crc kubenswrapper[4926]: I0312 18:42:56.818780 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:43:26 crc kubenswrapper[4926]: I0312 18:43:26.818079 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:43:26 crc kubenswrapper[4926]: I0312 18:43:26.818746 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:43:56 crc kubenswrapper[4926]: I0312 18:43:56.818015 4926 patch_prober.go:28] interesting pod/machine-config-daemon-hmdg8 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 18:43:56 crc kubenswrapper[4926]: I0312 18:43:56.818538 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 18:43:56 crc kubenswrapper[4926]: I0312 18:43:56.818578 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" Mar 12 18:43:56 crc kubenswrapper[4926]: I0312 18:43:56.819045 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2"} pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 18:43:56 crc kubenswrapper[4926]: I0312 18:43:56.819090 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerName="machine-config-daemon" containerID="cri-o://79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2" gracePeriod=600 Mar 12 18:43:56 crc kubenswrapper[4926]: E0312 18:43:56.957582 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:43:57 crc kubenswrapper[4926]: I0312 18:43:57.360858 4926 generic.go:334] "Generic (PLEG): container finished" podID="f7b34559-da2f-4796-8f3f-c56b2725c464" containerID="79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2" exitCode=0 Mar 12 18:43:57 crc kubenswrapper[4926]: I0312 18:43:57.360903 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" event={"ID":"f7b34559-da2f-4796-8f3f-c56b2725c464","Type":"ContainerDied","Data":"79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2"} Mar 12 18:43:57 crc kubenswrapper[4926]: I0312 18:43:57.360960 4926 scope.go:117] "RemoveContainer" containerID="2201cfc89c392b2d9df343a46350dae7f7620675e753c895f182c00c5eb9467c" Mar 12 18:43:57 crc kubenswrapper[4926]: I0312 18:43:57.361787 4926 scope.go:117] "RemoveContainer" containerID="79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2" Mar 12 18:43:57 crc kubenswrapper[4926]: E0312 18:43:57.362270 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.150468 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29555684-njhqk"] Mar 12 18:44:00 crc kubenswrapper[4926]: E0312 18:44:00.151347 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2955f1b9-7c06-4681-b024-65ce46309648" containerName="oc" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.151364 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2955f1b9-7c06-4681-b024-65ce46309648" containerName="oc" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.151646 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2955f1b9-7c06-4681-b024-65ce46309648" containerName="oc" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.152350 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555684-njhqk" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.155118 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.156099 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.156568 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-24cm5" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.159628 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555684-njhqk"] Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.206676 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlw9x\" (UniqueName: \"kubernetes.io/projected/6291e99f-3a9d-4fa6-a390-7000bc4bb107-kube-api-access-hlw9x\") pod \"auto-csr-approver-29555684-njhqk\" (UID: \"6291e99f-3a9d-4fa6-a390-7000bc4bb107\") " pod="openshift-infra/auto-csr-approver-29555684-njhqk" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.308126 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlw9x\" (UniqueName: \"kubernetes.io/projected/6291e99f-3a9d-4fa6-a390-7000bc4bb107-kube-api-access-hlw9x\") pod \"auto-csr-approver-29555684-njhqk\" (UID: \"6291e99f-3a9d-4fa6-a390-7000bc4bb107\") " pod="openshift-infra/auto-csr-approver-29555684-njhqk" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.329682 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlw9x\" (UniqueName: \"kubernetes.io/projected/6291e99f-3a9d-4fa6-a390-7000bc4bb107-kube-api-access-hlw9x\") pod \"auto-csr-approver-29555684-njhqk\" (UID: \"6291e99f-3a9d-4fa6-a390-7000bc4bb107\") " pod="openshift-infra/auto-csr-approver-29555684-njhqk" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.475234 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555684-njhqk" Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.960026 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29555684-njhqk"] Mar 12 18:44:00 crc kubenswrapper[4926]: I0312 18:44:00.966378 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 18:44:01 crc kubenswrapper[4926]: I0312 18:44:01.405611 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555684-njhqk" event={"ID":"6291e99f-3a9d-4fa6-a390-7000bc4bb107","Type":"ContainerStarted","Data":"64143db8c03f28b46b904bc03153011d44e646babf188aa5c3447de2a0718546"} Mar 12 18:44:01 crc kubenswrapper[4926]: E0312 18:44:01.422480 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = copying system image from manifest list: determining manifest MIME type for docker://registry.redhat.io/openshift4/ose-cli:latest: Manifest does not match provided manifest digest sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 12 18:44:01 crc kubenswrapper[4926]: E0312 18:44:01.422721 4926 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 12 18:44:01 crc kubenswrapper[4926]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 12 18:44:01 crc kubenswrapper[4926]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hlw9x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29555684-njhqk_openshift-infra(6291e99f-3a9d-4fa6-a390-7000bc4bb107): ErrImagePull: copying system image from manifest list: determining manifest MIME type for docker://registry.redhat.io/openshift4/ose-cli:latest: Manifest does not match provided manifest digest sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9 Mar 12 18:44:01 crc kubenswrapper[4926]: > logger="UnhandledError" Mar 12 18:44:01 crc kubenswrapper[4926]: E0312 18:44:01.424352 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"copying system image from manifest list: determining manifest MIME type for docker://registry.redhat.io/openshift4/ose-cli:latest: Manifest does not match provided manifest digest sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\"" pod="openshift-infra/auto-csr-approver-29555684-njhqk" podUID="6291e99f-3a9d-4fa6-a390-7000bc4bb107" Mar 12 18:44:02 crc kubenswrapper[4926]: E0312 18:44:02.415462 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29555684-njhqk" podUID="6291e99f-3a9d-4fa6-a390-7000bc4bb107" Mar 12 18:44:09 crc kubenswrapper[4926]: I0312 18:44:09.490338 4926 scope.go:117] "RemoveContainer" containerID="79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2" Mar 12 18:44:09 crc kubenswrapper[4926]: E0312 18:44:09.491477 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:44:18 crc kubenswrapper[4926]: I0312 18:44:18.590714 4926 generic.go:334] "Generic (PLEG): container finished" podID="6291e99f-3a9d-4fa6-a390-7000bc4bb107" containerID="0b3f89f6419767ede95419fb393039cf31985861b56ae23c4e99cb6ee9508a18" exitCode=0 Mar 12 18:44:18 crc kubenswrapper[4926]: I0312 18:44:18.590857 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555684-njhqk" event={"ID":"6291e99f-3a9d-4fa6-a390-7000bc4bb107","Type":"ContainerDied","Data":"0b3f89f6419767ede95419fb393039cf31985861b56ae23c4e99cb6ee9508a18"} Mar 12 18:44:19 crc kubenswrapper[4926]: I0312 18:44:19.945315 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555684-njhqk" Mar 12 18:44:20 crc kubenswrapper[4926]: I0312 18:44:20.106349 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlw9x\" (UniqueName: \"kubernetes.io/projected/6291e99f-3a9d-4fa6-a390-7000bc4bb107-kube-api-access-hlw9x\") pod \"6291e99f-3a9d-4fa6-a390-7000bc4bb107\" (UID: \"6291e99f-3a9d-4fa6-a390-7000bc4bb107\") " Mar 12 18:44:20 crc kubenswrapper[4926]: I0312 18:44:20.116773 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6291e99f-3a9d-4fa6-a390-7000bc4bb107-kube-api-access-hlw9x" (OuterVolumeSpecName: "kube-api-access-hlw9x") pod "6291e99f-3a9d-4fa6-a390-7000bc4bb107" (UID: "6291e99f-3a9d-4fa6-a390-7000bc4bb107"). InnerVolumeSpecName "kube-api-access-hlw9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:44:20 crc kubenswrapper[4926]: I0312 18:44:20.208272 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlw9x\" (UniqueName: \"kubernetes.io/projected/6291e99f-3a9d-4fa6-a390-7000bc4bb107-kube-api-access-hlw9x\") on node \"crc\" DevicePath \"\"" Mar 12 18:44:20 crc kubenswrapper[4926]: I0312 18:44:20.496106 4926 scope.go:117] "RemoveContainer" containerID="79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2" Mar 12 18:44:20 crc kubenswrapper[4926]: E0312 18:44:20.496681 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:44:20 crc kubenswrapper[4926]: I0312 18:44:20.610747 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29555684-njhqk" Mar 12 18:44:20 crc kubenswrapper[4926]: I0312 18:44:20.610717 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29555684-njhqk" event={"ID":"6291e99f-3a9d-4fa6-a390-7000bc4bb107","Type":"ContainerDied","Data":"64143db8c03f28b46b904bc03153011d44e646babf188aa5c3447de2a0718546"} Mar 12 18:44:20 crc kubenswrapper[4926]: I0312 18:44:20.610867 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64143db8c03f28b46b904bc03153011d44e646babf188aa5c3447de2a0718546" Mar 12 18:44:21 crc kubenswrapper[4926]: I0312 18:44:21.016738 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29555678-f2ldc"] Mar 12 18:44:21 crc kubenswrapper[4926]: I0312 18:44:21.025077 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29555678-f2ldc"] Mar 12 18:44:22 crc kubenswrapper[4926]: I0312 18:44:22.502394 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e31677c-4037-4cbc-9d34-5487431793e9" path="/var/lib/kubelet/pods/6e31677c-4037-4cbc-9d34-5487431793e9/volumes" Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.754556 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nmw49"] Mar 12 18:44:24 crc kubenswrapper[4926]: E0312 18:44:24.754949 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6291e99f-3a9d-4fa6-a390-7000bc4bb107" containerName="oc" Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.754963 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6291e99f-3a9d-4fa6-a390-7000bc4bb107" containerName="oc" Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.755240 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6291e99f-3a9d-4fa6-a390-7000bc4bb107" containerName="oc" Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.757476 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.773200 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nmw49"] Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.902397 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frg9g\" (UniqueName: \"kubernetes.io/projected/6a674ded-e154-446d-a265-811bd73ca4fe-kube-api-access-frg9g\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.902550 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-catalog-content\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:24 crc kubenswrapper[4926]: I0312 18:44:24.902658 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-utilities\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.004630 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frg9g\" (UniqueName: \"kubernetes.io/projected/6a674ded-e154-446d-a265-811bd73ca4fe-kube-api-access-frg9g\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.004716 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-catalog-content\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.004796 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-utilities\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.005299 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-utilities\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.005687 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-catalog-content\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.029147 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frg9g\" (UniqueName: \"kubernetes.io/projected/6a674ded-e154-446d-a265-811bd73ca4fe-kube-api-access-frg9g\") pod \"community-operators-nmw49\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.090596 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.595939 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nmw49"] Mar 12 18:44:25 crc kubenswrapper[4926]: I0312 18:44:25.663780 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmw49" event={"ID":"6a674ded-e154-446d-a265-811bd73ca4fe","Type":"ContainerStarted","Data":"fef7a38b7146840adf84cbcc717a5e2dfa9ef3dbdfc1495581961ca7a988f5bb"} Mar 12 18:44:26 crc kubenswrapper[4926]: I0312 18:44:26.708412 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a674ded-e154-446d-a265-811bd73ca4fe" containerID="479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0" exitCode=0 Mar 12 18:44:26 crc kubenswrapper[4926]: I0312 18:44:26.708498 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmw49" event={"ID":"6a674ded-e154-446d-a265-811bd73ca4fe","Type":"ContainerDied","Data":"479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0"} Mar 12 18:44:28 crc kubenswrapper[4926]: I0312 18:44:28.752037 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a674ded-e154-446d-a265-811bd73ca4fe" containerID="8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8" exitCode=0 Mar 12 18:44:28 crc kubenswrapper[4926]: I0312 18:44:28.752092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmw49" event={"ID":"6a674ded-e154-446d-a265-811bd73ca4fe","Type":"ContainerDied","Data":"8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8"} Mar 12 18:44:29 crc kubenswrapper[4926]: I0312 18:44:29.763595 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmw49" event={"ID":"6a674ded-e154-446d-a265-811bd73ca4fe","Type":"ContainerStarted","Data":"8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0"} Mar 12 18:44:33 crc kubenswrapper[4926]: I0312 18:44:33.489561 4926 scope.go:117] "RemoveContainer" containerID="79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2" Mar 12 18:44:33 crc kubenswrapper[4926]: E0312 18:44:33.491584 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" Mar 12 18:44:35 crc kubenswrapper[4926]: I0312 18:44:35.091961 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:35 crc kubenswrapper[4926]: I0312 18:44:35.092295 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:35 crc kubenswrapper[4926]: I0312 18:44:35.150198 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:35 crc kubenswrapper[4926]: I0312 18:44:35.170782 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nmw49" podStartSLOduration=8.68482412 podStartE2EDuration="11.170762056s" podCreationTimestamp="2026-03-12 18:44:24 +0000 UTC" firstStartedPulling="2026-03-12 18:44:26.714106494 +0000 UTC m=+2507.082732827" lastFinishedPulling="2026-03-12 18:44:29.2000444 +0000 UTC m=+2509.568670763" observedRunningTime="2026-03-12 18:44:29.783779449 +0000 UTC m=+2510.152405782" watchObservedRunningTime="2026-03-12 18:44:35.170762056 +0000 UTC m=+2515.539388389" Mar 12 18:44:35 crc kubenswrapper[4926]: I0312 18:44:35.865283 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:35 crc kubenswrapper[4926]: I0312 18:44:35.931902 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nmw49"] Mar 12 18:44:37 crc kubenswrapper[4926]: I0312 18:44:37.838662 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nmw49" podUID="6a674ded-e154-446d-a265-811bd73ca4fe" containerName="registry-server" containerID="cri-o://8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0" gracePeriod=2 Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.302029 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.377770 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frg9g\" (UniqueName: \"kubernetes.io/projected/6a674ded-e154-446d-a265-811bd73ca4fe-kube-api-access-frg9g\") pod \"6a674ded-e154-446d-a265-811bd73ca4fe\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.377902 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-utilities\") pod \"6a674ded-e154-446d-a265-811bd73ca4fe\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.377929 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-catalog-content\") pod \"6a674ded-e154-446d-a265-811bd73ca4fe\" (UID: \"6a674ded-e154-446d-a265-811bd73ca4fe\") " Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.379027 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-utilities" (OuterVolumeSpecName: "utilities") pod "6a674ded-e154-446d-a265-811bd73ca4fe" (UID: "6a674ded-e154-446d-a265-811bd73ca4fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.379879 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.390656 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a674ded-e154-446d-a265-811bd73ca4fe-kube-api-access-frg9g" (OuterVolumeSpecName: "kube-api-access-frg9g") pod "6a674ded-e154-446d-a265-811bd73ca4fe" (UID: "6a674ded-e154-446d-a265-811bd73ca4fe"). InnerVolumeSpecName "kube-api-access-frg9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.482056 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frg9g\" (UniqueName: \"kubernetes.io/projected/6a674ded-e154-446d-a265-811bd73ca4fe-kube-api-access-frg9g\") on node \"crc\" DevicePath \"\"" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.642747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a674ded-e154-446d-a265-811bd73ca4fe" (UID: "6a674ded-e154-446d-a265-811bd73ca4fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.686854 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a674ded-e154-446d-a265-811bd73ca4fe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.851324 4926 generic.go:334] "Generic (PLEG): container finished" podID="6a674ded-e154-446d-a265-811bd73ca4fe" containerID="8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0" exitCode=0 Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.851372 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmw49" event={"ID":"6a674ded-e154-446d-a265-811bd73ca4fe","Type":"ContainerDied","Data":"8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0"} Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.851401 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nmw49" event={"ID":"6a674ded-e154-446d-a265-811bd73ca4fe","Type":"ContainerDied","Data":"fef7a38b7146840adf84cbcc717a5e2dfa9ef3dbdfc1495581961ca7a988f5bb"} Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.851401 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nmw49" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.851420 4926 scope.go:117] "RemoveContainer" containerID="8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.899313 4926 scope.go:117] "RemoveContainer" containerID="8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.904456 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nmw49"] Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.910806 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nmw49"] Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.934310 4926 scope.go:117] "RemoveContainer" containerID="479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.971683 4926 scope.go:117] "RemoveContainer" containerID="8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0" Mar 12 18:44:38 crc kubenswrapper[4926]: E0312 18:44:38.972852 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0\": container with ID starting with 8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0 not found: ID does not exist" containerID="8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.972900 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0"} err="failed to get container status \"8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0\": rpc error: code = NotFound desc = could not find container \"8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0\": container with ID starting with 8e66a155454b6e73008ebb3a98d0d518ae2966bfdbbe486df832f6aac00d97b0 not found: ID does not exist" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.972929 4926 scope.go:117] "RemoveContainer" containerID="8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8" Mar 12 18:44:38 crc kubenswrapper[4926]: E0312 18:44:38.975186 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8\": container with ID starting with 8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8 not found: ID does not exist" containerID="8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.975215 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8"} err="failed to get container status \"8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8\": rpc error: code = NotFound desc = could not find container \"8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8\": container with ID starting with 8374c99f44372f42f5951b3b9b5caeee7c593394cc4227291be1f3a6c4e9c2a8 not found: ID does not exist" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.975233 4926 scope.go:117] "RemoveContainer" containerID="479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0" Mar 12 18:44:38 crc kubenswrapper[4926]: E0312 18:44:38.975612 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0\": container with ID starting with 479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0 not found: ID does not exist" containerID="479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0" Mar 12 18:44:38 crc kubenswrapper[4926]: I0312 18:44:38.975636 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0"} err="failed to get container status \"479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0\": rpc error: code = NotFound desc = could not find container \"479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0\": container with ID starting with 479a94d3e549210bdc35aced5c4bf42a18c01316139fb6e19becde88f2ce67f0 not found: ID does not exist" Mar 12 18:44:40 crc kubenswrapper[4926]: I0312 18:44:40.510329 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a674ded-e154-446d-a265-811bd73ca4fe" path="/var/lib/kubelet/pods/6a674ded-e154-446d-a265-811bd73ca4fe/volumes" Mar 12 18:44:44 crc kubenswrapper[4926]: I0312 18:44:44.490712 4926 scope.go:117] "RemoveContainer" containerID="79944b20ee2855a4a0a1dba29df2c15482bb6976a2ebdd8437759ebdec65cbd2" Mar 12 18:44:44 crc kubenswrapper[4926]: E0312 18:44:44.491339 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-hmdg8_openshift-machine-config-operator(f7b34559-da2f-4796-8f3f-c56b2725c464)\"" pod="openshift-machine-config-operator/machine-config-daemon-hmdg8" podUID="f7b34559-da2f-4796-8f3f-c56b2725c464" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515154604655024460 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015154604656017376 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015154577340016520 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015154577340015470 5ustar corecore